Reference list for online contents#
The table below lists commands, scripts, and figures included in the book and contained in the online compendium: accessing the online version of each content can be achieved by clicking on the relevant ID
field.
Elements are organised alphabetically according to the ID
value. Clicking on a column header changes the sorting order; the included Search field looks for a string across all contents (i.e. the search is applied to all columns at once).
ID |
description |
page |
---|---|---|
Default appearance of terminal when |
101 |
|
Command to create a new virtual environment in |
101 |
|
Command to activate a virtual environment in |
101 |
|
Command to deactivate the virtual environment in |
101 |
|
Default appearance of terminal when |
101 |
|
Command to install |
101 |
|
Initiate |
124 |
|
Clone a remote repository |
124 |
|
Add all changes (even from previously untracked files) to the local |
125 |
|
Record ( |
125 |
|
Send ( |
125 |
|
Obtain ( |
125 |
|
Include/apply ( |
125 |
|
Obtain and include/apply ( |
125 |
|
Install |
152 |
|
Setup |
153 |
|
Start |
153 |
|
Install |
156 |
|
Start |
156 |
|
Start |
156 |
|
Install package |
157 |
|
Use |
157 |
|
Use |
157 |
|
Example of the XML structure created by |
158-159 |
|
Use |
159 |
|
Extract links from HTML pages using |
162-163 |
|
Download and scrape HTML pages from links extracted with |
164-166 |
|
Download and scrape HTML pages from links extracted with |
166 |
|
Extract metadata from the downloaded HTML pages using |
166-171 |
|
Basic structure of the metadata table included in MoreThesis pages |
171-173 |
|
Download PDF files linked in HTML pages |
174-175 |
|
Extract the contents of PDF files as plain-text using |
176 |
|
Create an XML corpus combining the metadata from HTML pages and the contents of PDF files using |
177-180 |
|
Install |
183 |
|
Basic |
183 |
|
Accessing the “help” section for a specific |
183 |
|
Basic syntax to scrape tweets to .jsonl using |
191 |
|
Example of how to scrape tweets to .jsonl using |
191 |
|
Example of |
192 |
|
Example of |
192 |
|
Example of |
192 |
|
Example of |
192 |
|
Example of |
192 |
|
Install |
193 |
|
Use script |
193 |
|
Scrape tweets with |
193-196 |
|
Example of filename saved by script |
196 |
|
Metadata data points collected by |
197-203 |
|
Example of data extracted with |
204 |
|
Convert tweets extracted with |
204-206 |
|
Install |
206 |
|
Basic |
206 |
|
Example of |
208 |
|
Metadata data points collected by |
209-218 |
|
Metadata data points collected by |
218 |
|
Example of data extracted with |
220 |
|
Convert Instagram posts and comments extracted with |
220-226 |
|
Install |
228 |
|
Basic |
228 |
|
Metadata data points collected by |
229-233 |
|
Metadata data points collected by |
233-235 |
|
Metadata data points collected by |
236 |
|
Example of data extracted with |
236-237 |
|
Convert Facebook posts and comments extracted with |
237-242 |
|
Get profile details from Facebook using |
242-245 |
|
Implement the collection of profile details ( |
245-246 |
|
Install |
247 |
|
Basic |
247 |
|
Example of TTML format |
252-253 |
|
Example of SRV format without auto-captioning |
253 |
|
Example of SRV format with auto-captioning |
253-254 |
|
Install |
254 |
|
Basic |
254 |
|
Metadata data points collected by |
255-262 |
|
Metadata data points collected by |
262 |
|
Extracting video details, metadata, and subtitles from Youtube without multimedia files |
263 |
|
Example of data extracted with |
264 |
|
Example of data extracted with |
264 |
|
Extract collected Youtube data (everything except comments) to XML format |
264-269 |
|
Extract collected Youtube comments to XML format |
269-272 |
|
Sample usage of the module |
274 |
|
Example of recognised spelling variants in VARD |
276 |
|
Example of unrecognised spelling variants in VARD |
277 |
|
Example of normalised data in XML format generated with VARD |
278 |
|
Install |
279 |
|
Basic |
279 |
|
Identify a set of predefined languages in .txt files and write a summary report in spreadsheet format |
281-283 |
|
Example of hashtags transformed through |
286 |
|
Segment hashtags and transform them into XML tags in a XML corpus file |
287-288 |
|
Regular expression to capture usernames/username handles |
289 |
|
Regular expression to capture simple URLs |
289 |
|
Regular expression to capture complex URLs |
289 |
|
Regular expression to capture cashtags |
289 |
|
Install |
291 |
|
Install |
291 |
|
Example of data in XML format extracted with |
292 |
|
Annotate |
292-294 |
|
Annotate |
294-295 |
|
Example of XML verticalised format |
296 |
|
OpenRefine main page |
298 |
|
Preview for CSV import in OpenRefine |
299 |
|
Preview for JSON import in OpenRefine |
299 |
|
Preview for XML import – step 1 – in OpenRefine |
300 |
|
Preview for XML import – step 2 – in OpenRefine |
300 |
|
Using ‘facets’ (filters) in OpenRefine |
301 |
|
Example of a page collected from the Silk Road 1 forum |
317 |
|
Example (modified) of the post structure in Silk Road 1 HTML pages |
320-321 |
|
XML meta-structure of the data extracted through |
321-322 |
|
Convert Silk Road 1 HTML pages to XML format using |
323-328 |
|
XML meta-structure of the documents included in the DPM corpus |
328 |
|
Scrape tweets created after a specific date with |
337 |
|
Scrape tweets created after a specific date with |
338 |
|
Convert tweets extracted with |
338-340 |
|
Example of data extracted with |
340 |
|
Example of syntax used by WordPress to show all posts available in a website |
342 |
|
Collect (crawl) all posts links from a WordPress website |
342-344 |
|
Example of a message containing an emoji |
344 |
|
Examples of emoji transliterations applied to |
344-345 |
|
Function to transliterate emojis using the |
345-346 |
|
Example of data extracted with |
352-353 |
|
Collect all chatlogs from perverted-justice.com |
353-357 |
|
Convert PJ chatlogs into XML format |
360-366 |
|
Example of the interactive plot created for the visual exploration of collocations |
368 |
|
#LancsBox main interface |
149 |
|
#LancsBox data collection interface |
149 |