Cara Mendapatkan Data (Crawl) Twitter X - 30 Maret 2024

Helmi Satria
30 Mar 202418:04

Summary

TLDRThis video provides an updated tutorial on how to crawl data from Twitter (or X) using the custom-made tool 'Tweet Harvest'. It explains how to set up the tool via Google Colab, acquire Twitter authentication tokens, and gather data based on specific keywords. The creator also addresses common issues with data collection, including rate limits, and gives tips for overcoming these challenges. The tool is intended for research purposes only, and viewers are advised not to use it for commercial activities. The video concludes by guiding viewers on how to save and process the collected data.

Takeaways

  • 🔍 To crawl Twitter/X data based on a specific search keyword, the tool Tweet Harvest can be used.
  • đŸš« If the previous method for crawling data is not working, updating to the latest version (v2.60) of Tweet Harvest should resolve the issue.
  • ⏳ Due to Twitter's rate limits, you can only retrieve about 500 tweets every 10-15 minutes. Over a full day, this could result in collecting around 2,700 tweets, but manual stopping is recommended to avoid account blocking.
  • 🔑 To use Tweet Harvest, you need to obtain a Twitter OAuth token, which acts like a password. This can be done by inspecting the Twitter web page's cookies after logging in.
  • ⚙ Tweet Harvest is built using Node.js, and requires certain installations, such as Pandas and Node.js, to work on platforms like Google Colab.
  • 📝 The CSV output file will contain detailed tweet data, including timestamps, tweet content, likes, and usernames, which can be accessed and analyzed further in tools like Excel or Google Sheets.
  • 📊 While the Tweet Harvest tool can gather data for research purposes, it should not be used for business or monetization purposes.
  • 🗂 The keyword, language, and number of tweets to be crawled can be customized, allowing for various types of research (e.g., analyzing tweets related to specific events or people).
  • 📅 Data can be filtered by time range, and you can search tweets from specific users with specific content (e.g., tweets mentioning COVID-19 from a particular public figure).
  • ⛔ If the crawler hits a rate limit, it will pause for 10 minutes before continuing. Errors are expected, but they do not prevent the tool from working.

Q & A

  • What is Tweet Harvest, and how does it work?

    -Tweet Harvest is a tool created by the speaker for crawling data from Twitter (now X). It collects tweets based on specific search keywords. It uses Node.js for its backend and scrolls through Twitter to gather data such as tweet text, likes, replies, and more. The tool outputs the data into a CSV file for further analysis.

  • Why can't some users get Tweet Harvest to work?

    -Users may experience issues with Tweet Harvest if they're using outdated versions of the tool. The speaker recommends using the latest stable version (260 as of March 30) for optimal performance. Additionally, issues could stem from incorrect setup or limits imposed by Twitter’s API.

  • What are the rate limits imposed by Twitter for data collection using Tweet Harvest?

    -Twitter imposes rate limits that restrict the number of tweets that can be collected within a given time frame. The speaker mentions a limit of about 500 tweets every 10-15 minutes. Users need to wait for these limits to reset before continuing to collect data.

  • How much data can be collected in a day using Tweet Harvest?

    -The speaker states that they have been able to collect around 2,700 tweets in a day, but this number is not fixed. It largely depends on how long the user is willing to wait for rate limits to reset, as well as how long the tool is allowed to run.

  • How do you set up Tweet Harvest for a new project?

    -To set up Tweet Harvest, users need to obtain their Twitter auth token, which serves as a kind of password for accessing their account’s API. After obtaining the token, users input their search keyword, desired tweet language, and other parameters into the tool before running it to collect data.

  • What are some examples of keywords that can be used in Tweet Harvest?

    -Users can input any keyword relevant to their research. For example, the speaker mentions collecting tweets about Indonesian President Jokowi by setting the keyword as 'Jokowi' and setting the language to Indonesian ('lang: id'). Other keywords could include topics like elections, COVID, or political figures.

  • What kind of data does Tweet Harvest collect?

    -Tweet Harvest collects a variety of data from Twitter, including the tweet text, number of likes, replies, retweets, the time of creation, whether an image is attached, the username of the person who tweeted, and the tweet's URL. All this data is stored in a CSV file for further processing.

  • How does the tool handle rate limits when collecting a large number of tweets?

    -Tweet Harvest automatically stops collecting tweets once it hits Twitter’s rate limit (around 500 tweets per 10-15 minutes). After the rate limit is reached, the tool waits for 10 minutes before continuing to scroll and collect more data.

  • Can the tool be used for purposes other than research?

    -The speaker explicitly advises against using Tweet Harvest for commercial purposes or any activities beyond research, such as thesis work or dissertation. The tool was built specifically to aid academic research and should not be used for business or profit-driven activities.

  • How can users further process the collected Twitter data?

    -After collecting the data in a CSV file, users can process it further using tools like Excel, Google Sheets, or even programming languages like Python. They can perform sentiment analysis, social network analysis, or other forms of data analysis based on their needs.

Outlines

00:00

🐩 Quick Update on Twitter/X Data Crawling

In this paragraph, the speaker gives a quick update about how to crawl data from Twitter (now known as X). They mention that previous methods may not work anymore and address common questions about the limitations of data retrieval. The speaker shares that using Tweet Harvest, a tool they built, users can gather about 500 pieces of data every 10 to 15 minutes, and up to around 2,700 in a day. This is mainly for academic purposes like research, theses, or dissertations, and not for commercial use. They emphasize that users should not exploit this tool for non-research purposes.

05:00

đŸ› ïž Setting Up and Running Tweet Harvest with Google Colab

This section explains how to use Tweet Harvest through Google Colab. It walks the user through the steps to copy the tool, access Twitter’s authentication token, and how to safely share and handle it. The speaker demonstrates how to install necessary packages, such as Pandas and Node.js, and how the Tweet Harvest tool crawls data by scrolling and capturing Twitter posts. Users are guided on how to customize file names and search keywords to save their data in .csv format.

10:03

🔧 Detailed Steps for Using Tweet Harvest Tool

The speaker continues to explain the process of using Tweet Harvest, focusing on the installation of Node.js and setting specific search parameters. They use an example of crawling data related to President Jokowi and demonstrate how to filter tweets based on the chosen search keywords. The paragraph details how the tool saves the crawled data, handles Twitter rate limits, and what to expect when running the script multiple times, including how to manage and download CSV files.

15:03

📊 Working with the Crawled Twitter Data

This paragraph covers the process of analyzing and working with the data gathered from Twitter. The speaker talks about how the crawled data, which includes information like tweet text, likes, retweets, and usernames, can be exported and accessed through CSV files. They also discuss the rate limit imposed by Twitter, how to work around it by waiting, and potential errors that may arise after collecting around 500–600 tweets. The speaker emphasizes that patience is required to gather more data.

💡 Additional Tips and Advanced Data Crawling

In this final section, the speaker answers some remaining questions about crawling specific types of data, such as tweets within a certain time range, tweets by specific users, or tweets containing minimum likes and replies. They mention that these methods can be found in more detail in their blog and offer additional tools for advanced data analysis, such as sentiment analysis. The speaker concludes by encouraging viewers to check their blog for further guidance and details on preprocessing and using the collected Twitter data.

Mindmap

Keywords

💡Crawling Data

Data crawling refers to the process of programmatically gathering data from a website, in this case, Twitter (now X). This video explains how to use a tool named 'tweet Harvest' to collect tweets based on certain search keywords, especially for academic purposes like research and analysis.

💡Tweet Harvest

Tweet Harvest is a custom-built tool used in the video to crawl Twitter data. It allows users to set parameters, such as specific keywords and time frames, to collect tweet data. The creator of the video uses it to demonstrate safe and ethical data collection methods for research.

💡Search Keyword

Search keywords are terms entered to filter tweets on specific topics. The video emphasizes using keywords, such as 'Jokowi' or 'Pemilu,' to retrieve relevant tweet data, illustrating how to refine searches for better-targeted data collection in Tweet Harvest.

💡Rate Limiting

Rate limiting is a restriction set by Twitter to control the amount of data accessed within a certain timeframe. The video notes that Twitter allows around 500 tweets to be collected every 10-15 minutes, emphasizing patience in data gathering due to these limits.

💡Google Colab

Google Colab is a free cloud-based environment that supports Python and is used here to run Tweet Harvest. The video shows users how to set up and run Tweet Harvest in Google Colab, making the process accessible and easy for anyone without specialized software.

💡OAuth Token

An OAuth token is a security credential that authorizes access to Twitter’s data. In the video, the speaker explains how to obtain this token from Twitter, as it's essential to access the data using Tweet Harvest securely and avoid unauthorized usage.

💡CSV Output

CSV output refers to the data format that Tweet Harvest exports. The video shows that after collecting data, the results are saved in a CSV file (e.g., 'jokowi.csv'), which can be opened and analyzed in spreadsheet software like Excel or Google Sheets.

💡Rate Limit Errors

Rate limit errors occur when the data collection surpasses Twitter’s set limits. The video addresses this by reassuring viewers that such errors are normal when Tweet Harvest reaches around 500 tweets, and advises users to wait for the rate limit to reset before resuming.

💡Sentiment Analysis

Sentiment analysis is the process of analyzing tweets for their positive or negative sentiment. The video implies that, once the tweets are collected, they can be used for sentiment analysis to gauge public opinion on topics such as elections or political figures like 'Jokowi.'

💡Language Filter (lang: id)

The language filter 'lang: id' is used to collect tweets written in Indonesian. This feature in Tweet Harvest helps researchers collect data specific to a language, allowing them to filter for region-specific content for more relevant analysis.

Highlights

Quick update on obtaining Twitter data using Tweet Harvest, addressing recent user issues with prior versions.

Overview of Tweet Harvest functionality, which allows Twitter data crawling based on keywords with token authentication.

Detailed instructions for generating and managing an OAuth token needed to access Twitter data.

Explanation of rate limits on data collection: around 500 tweets every 10–15 minutes, reaching about 2,700 tweets daily.

A reminder that Tweet Harvest is for research purposes only, specifically for theses, dissertations, or academic studies.

Guide to using Google Colab with Tweet Harvest to simplify the data collection process and create editable copies in Google Drive.

Step-by-step tutorial on setting up Google Colab, including essential tools like Pandas and Node.js installation for Tweet Harvest.

How to specify the keyword search criteria and store data in a CSV file with customizable names, e.g., 'jokowi.csv.'

Clarification on the use of the language filter (e.g., 'lang: id') to collect tweets in specific languages such as Indonesian.

Explanation of error handling during crawling due to rate limits; waiting 10 minutes before continuing to collect more data.

Tips for extracting and storing tweet data, including metadata like tweet timestamp, likes, images, username, and location.

Instructions for downloading and using the collected CSV file offline or in programs like Excel and Google Sheets.

Advice on using Tweet Harvest for sentiment analysis or other analyses once data collection is complete.

Brief overview of Twitter API limitations that prevent many researchers from collecting data, with Tweet Harvest as a workaround.

Additional details on specifying search parameters such as exact phrases, minimum likes or replies, and timeframes for targeted data collection.

Transcripts

play00:00

cara mendapatkan data atau craowl data

play00:02

dari platform Twitter atau X ini aku mau

play00:06

quick update aja dari aku kan udah

play00:09

beberapa video terakhir kan bikin cara

play00:12

mendapatkan data Twitter terus ya tapi

play00:14

ini aku quick update aja soalnya banyak

play00:16

yang nanyain Kenapa pada enggak bisa

play00:19

jadi di video ini aku cuma mau bahas

play00:22

tiga pertanyaan ini yaitu ya cara cara

play00:26

craowl data Twitter berdaskan search

play00:28

keyword tertentu itu caranya gimana nah

play00:30

terus kok yang sebelumnya enggak bisa

play00:32

mas gitu aku udah pakai twet Harvest

play00:34

yang versi 228 yang versi berapa versi

play00:37

berapa itu kok enggak bisa terus gimana

play00:39

gitu sama sebenarnya kita bisa berapa

play00:43

data sih yang bisa kita dapatin

play00:46

gitu short tanser-nya ini sebenarnya

play00:48

kita bisa dapetin Aku enggak tahu sih

play00:51

dapetin berapa cuman limitasinya itu per

play00:54

10 menit atau 10 sampai 15 menit itu tuh

play00:57

kita cuman dapetin bisa 500-an

play01:01

gitu Jadi per 10 atau 10 sampai 15 menit

play01:04

itu tuh kita cuman bisa dapetin data

play01:08

sekitar 500-an bisa 500 400 600 kayak

play01:12

gitu Jadi per hari bisa dapat

play01:15

berapa yang sejauh ini aku sudah coba

play01:19

Itu aku bisa dapatin sekitar

play01:23

2.700-an gini ini cuman sampai 2.700

play01:27

karena aku stop ya tahuak enggak mau

play01:30

kebanyakan juga karena ini aku cuman

play01:32

buat nyobain dan buat bantuin

play01:35

teman-teman yang lagi skripsian aja

play01:37

sejauh ini ini harusnya udah aman untuk

play01:39

dapatin

play01:40

data dan quick reminder ini research

play01:44

proposes only ya teman-teman

play01:47

Jadi selain untuk riset selain untuk

play01:50

skripsi tesis disertasi pokoknya untuk

play01:53

studi Kalian please jangan pakai tweet

play01:56

Harvest yang aku bikin sendiri ini Jadi

play01:58

ini aku emang bikin sendiri

play02:00

dan ya untuk kebutuhan riset dan bantuin

play02:03

teman-teman aja gitu kalau buat bisnis

play02:06

buat cari duit

play02:08

buat apa ya apunlah pokoknya selain

play02:12

skripsian selain bikin tesis Please

play02:15

jangan pakai TW Harvest ini gitu ya

play02:19

Oke Langsung aja nanti aku bakal ngasih

play02:21

link di description jadi teman-teman

play02:24

bisa langsung klik klik klik klik

play02:26

duplicate aja nanti bisa di e buka link

play02:29

nya itu nanti kita bakal pakai Google

play02:32

collab buat memudahkan segalanya terus

play02:34

nanti teman-teman bisa ke file terus

play02:36

Save A copy in drive gitu ya biar

play02:40

teman-teman bisa dapatin Google Cab

play02:42

teman-teman sendiri gitu bisa diedit

play02:44

bisa di apaa terserah gitu Jadi ini

play02:49

langsung aja ya kita pertama-tama bakal

play02:51

butuh Twitter outoken

play02:53

karena alat yang aku bikin ini tweet

play02:56

Harvest itu tuh butuh

play02:58

autocun out Token itu apa itu kayak

play03:01

password password Twitter kalian jadi

play03:05

buat dapetin ininya Ini kan nanti

play03:07

digitnya banyak gini ya Nah itu tuh

play03:09

Caranya buka Twitter twitter.com login

play03:13

dulu kalau udah login kan halamannya

play03:15

bakal kayak gini nih terus ini diklik

play03:17

kanan terus

play03:19

inspect Klik Kanan inspect Nah nanti di

play03:22

bagian sini nanti ada icon yang kayak

play03:24

gini nih kalau teman-teman pakai

play03:26

Microsoft Edge Harusnya sama di Chrome

play03:28

pun harusnya application juga di

play03:30

application cookies ini di bagian sini

play03:33

nih nanti ada

play03:34

twitter.com habis itu ada name-nya out

play03:38

token gini out token nah value-nya ini

play03:42

kalian block terus dicopy semua nih copy

play03:47

gitu nah ini aku display ke kalian

play03:49

karena ini udah ku edit-edit ya Ini udah

play03:52

ku edit-edit jadi udah gak valid kalau

play03:54

teman-teman pakai jadi kalau teman-teman

play03:56

mau share aken atau Google colab kalian

play03:59

ke kalian ini diedit dulu ya nanti jadi

play04:03

udah dicopy kan udah dipaste gini nah

play04:05

ini nanti dihapus dulu sebelum kalian

play04:07

share gitu ya Soalnya ini kalian sama

play04:09

aja kayak share password kalian sendiri

play04:13

gitu gitu E ini udah dari yang tadi udah

play04:17

dicopy dipaste ke sini terus ini dihide

play04:20

aja habis itu ini kita run run run run

play04:24

aja ini buat instal

play04:26

Pandas habis itu instal Noe JS karena

play04:30

Harvest itu dibikin pakai Note JS jadi

play04:34

aku bikin pakai Note JS ya twet

play04:36

harvest-nya tweet Harvest itu apa sih

play04:39

cara kerjanya kayak gimana itu nanti

play04:41

bakal aku mungkin di di video yang lain

play04:45

soalnya bakal panjang intinya ini

play04:48

metodenya tuh craowling data gitu ya

play04:50

jadi dia tuh kayak

play04:52

eh scrollsroll sendiri gitu loh Scroll

play04:55

Scroll Scroll sendiri terus dapetin data

play04:57

Scroll dapetin data kayak gitu

play05:00

jadi manual Scroll

play05:01

aja gitu ya Nah udah instal Noe JS nanti

play05:06

bakal muncul versi nodejs yang keinstal

play05:08

yaitu versi sekitar 20-an kalau

play05:11

teman-teman lakuin ini setelah 30 Maret

play05:15

minimal versi 20-an ini habis note

play05:18

js-nya udah keinstal Harusnya kita udah

play05:20

bisa buat ngelakuin Crawl datanya untuk

play05:23

pakai tweet Harvest ini diun aja diklik

play05:27

play Oh ya ini file name-nya ya file

play05:31

nameennya ditentuin Kalian mau cral

play05:33

tentang apa misalkan tentang Pak Jokowi

play05:35

gitu ya Nah ini bisa diedit aja nanti

play05:37

bakal kesimpan di file name-nya itu

play05:40

jokowi.csv gitu Nah yang penting tuh

play05:43

Kalian ngedit search keyword-nya aja

play05:46

sesuai yang kalian butuhin mau tentang

play05:48

Jokowi tentang pemilu tentang Pak

play05:52

Prabowo siapun Apun yang kalian butuhin

play05:55

untuk reset Ini diganti-ganti aja ini

play06:00

titik du id untuk dapat bahasa

play06:03

Indonesia terus Kalian Butuh berapa data

play06:06

di sini aku 100 aja untuk demo

play06:08

proposes habis itu Line terakhir ini

play06:12

untuk jalanin tweet harvestnya Yang aku

play06:15

pakai yang 260 ya per 30 Maret ini yang

play06:19

versi terakhir yang paling Stable yang

play06:21

aku bisa dapatin sampai 2000an data ini

play06:24

pakai yang

play06:25

260 disimpan di jokowi.csv sesuai yang

play06:29

ini terus Search keyword-nya pakai

play06:31

Jokowi berbahasa Indonesia 100 data

play06:34

tokennya ambil dari yang atas tadi yang

play06:37

ada di sini ya

play06:40

udah habis itu diun harusnya bakal

play06:43

muncul kayak

play06:45

gini ya keep your access token

play06:48

Secret habis ituudah nih total twet save

play06:51

106 Pok kalau udah

play06:54

mencapai minimal itu limitnya Soalnya

play06:58

ini enggak bisa apa ya ya sebenarnya

play07:01

bisa aja sih buat exactly dapat 100 data

play07:04

gitu tapi ya karena tadi Scroll Scroll

play07:07

Scroll itu kan dia tiap kita Scroll tuh

play07:10

kita gak tahu Dapat berapa tweet kan

play07:12

dari Twitternya nah ini nambah-nambahnya

play07:14

tuh Pokoknya terserah Twitternya gitu

play07:16

Nah limitnya tadi itu batas minimalnya

play07:19

nah 100 itu n 106 Ini kan udah melebihi

play07:23

dari batas minimal Jadi udah stop dia

play07:26

got 106 twets dan scrolling gitu

play07:30

nah udah kita Run ini Nah itu tuh nanti

play07:34

dia bakal baca csv file yang udah

play07:37

kesimpan

play07:38

di di mana

play07:45

Sebentar cobat Refresh dulu

play07:55

ya Nah

play07:58

harusnya eh bentuk fil-ya kayak gini ya

play08:01

yang icon files ini kalau teman-teman

play08:03

klik dan udah ngejalanin croll datanya

play08:06

dan nanti itu hasilnya bakal kesimpan di

play08:08

twitch data terus tergantung kalian nama

play08:11

file name-nya apa tadi misalkan tadi kan

play08:14

gibran.csv gitu ya Nah itu tuh nanti

play08:17

bakal kesimpan di

play08:19

gibran.csv output yang penting

play08:21

sebenarnya ini Jadi kalian mau jalanin

play08:24

offline mau jalanin di Google colab mau

play08:26

jalanin di Jupiter notebook atau di mana

play08:29

Beb

play08:30

yang penting ini csv-nya jadi

play08:31

teman-teman kalau udah selesai di Google

play08:33

colab ini bisa diclose aja Google

play08:35

colab-nya tapi sebelum itu di-download

play08:37

dulu nih

play08:38

csv gitu ya Nah ini udah di-download

play08:43

harusnya Udah bisa diakses di Excel di

play08:46

Google spreadsheet harusnya udah aman

play08:48

sih gitu nah ini kolomnya udah banyak

play08:52

ada created dipost Kapan tweetnya berapa

play08:56

yang nglike

play08:58

berapa

play09:01

ada gambarnya atau

play09:02

enggak usernameennya siapa atau dia

play09:06

ngerely siapa lokasinya di

play09:10

mana terus ada tweet urlnya juga ada ada

play09:14

username yang ngweet juga ya harusnya

play09:16

udah cukup sih gitu ya jadi teman-teman

play09:19

bisa download csv-nya atau bisa ngproses

play09:22

lanjut di bawahnya

play09:24

lagi dan b Dar sebelnya ak bikinnya

play09:29

pakai titik koma sekarang udah pakai

play09:32

koma di mana itu bikin ini fnya k tutup

play09:37

dulu deh biar lebar Di mana itu bikin di

play09:40

Google colab tuh visualisasinya lebih

play09:43

cakep jadi di sini tu Kita

play09:46

bisaah ini kurun dulu deh

play09:49

coba nah habis diun kayak gini di kalau

play09:53

kita horizontal Scroll ke kanan di pojok

play09:56

kanan gini nah gitu di bagian atas pojok

play10:00

kanan atas tuh ada kayak gini nih nah

play10:03

itu kalau diklik tuh nah jadi bisa lebih

play10:04

rapi kan Nih ada 106 entries gituernya

play10:09

siapa tweet urlnya apa ini Harusnya ada

play10:12

gibran-gibrannya Nih nah ini ada Gibran

play10:15

tweet Gibran tweet di Gibran tweet

play10:19

e bahkan Gibran

play10:22

gitu ini kenapa ada

play10:28

gibrannya nah yang ini jujur ak gak tahu

play10:31

kenapa kenapa bisa muncul terus sini ada

play10:35

Gibran ya poknya gitu ya pokoknya ada

play10:38

gibran-gibrannya Harusnya nah ini Gibran

play10:42

gitu dan udah ada 106 harusnya udah oke

play10:46

udah

play10:47

aman harusnya ini ya Jadi nanti tadi

play10:51

yang sesuai ku jelasin

play10:54

waktu data itu setiap udah nyampai 500an

play10:58

dia itu harus nunggu dulu jadi karena

play11:01

ada rate limit dari

play11:02

Twitter yang 10 15 menitan tadi cuman

play11:06

bisa dapetin 500-an data Nah itu tuh

play11:08

nunggu 10 menit dan ini bakal auto Crow

play11:11

lagi habis nunggu 10 menit itu

play11:16

gitu ya kayak giniilah kurang lebih jadi

play11:20

Eh ini aku udah percobaan

play11:23

sebelumnya nah ini udah 300 400 500 nah

play11:28

di 600-an dia tuh kayak gini nih ada

play11:32

error n nah pokoknya kalau muncul kayak

play11:34

gini error kayak gini jangan kaget

play11:37

karena itu emang expected bakal muncul

play11:40

setiap teman-teman udah dapetin 600-an

play11:42

atau 500-an data ini nunggu 10 menitan

play11:46

habis 10 menitan nah ini bakal KW lagi

play11:50

gitu tuh udah 600an

play11:54

lagi kena rate limit lagi 10 menit

play11:57

berikutnya lanjut lagi dan gitu

play11:58

seterusnya

play12:01

gitu ya Nah terus ini limitasi dari

play12:05

Google aku ngakelimitasi Google clab-nya

play12:07

cuman buat dapatin data aja ya jadi

play12:10

setelah datanya udah dapat kalian mau

play12:12

apain Mau proses data mau sentimen

play12:14

analisis mau ngapain Terserah yang

play12:17

penting ini masalah hampir semua

play12:21

researcher di Indonesia atau di dunia

play12:23

sebenarnya itu gak bisaroll data karena

play12:26

api-nya kelimit kan jadi ya ini aku

play12:29

pengin Bantu teman-teman aja buat

play12:31

dapatin

play12:33

datanya gitu Jadi terserah nanti datanya

play12:36

habis itu kalian bisa lanjut di sini

play12:39

pakai data frame yang ini atau bisa

play12:41

download csv yang ada di folder twitch

play12:44

data ini diklik terus bisa di-download

play12:48

gitu terus bisa kalian pakai di mana

play12:52

aja itu harusnya ini udah menjawab Dua

play12:55

pertanyaan ya cara Craw data Twitter

play12:58

atau X berdasarkan search keyword

play13:00

tertentu terus kok cara sebelumnya itu

play13:03

gak bisa kenapa

play13:05

e Jujur aku engak

play13:08

tahu gak tahu sebenarnya ini kan dipakai

play13:10

di

play13:11

video-video content Creator yang lain ya

play13:13

di YouTube juga ada ada yang habis bikin

play13:15

konten e di beberapa bulan lalu juga ada

play13:18

yang bikin terus gak bisa juga dan

play13:21

harusnya semoga video ini bisa menjawab

play13:24

Gimana cara yang benarnya Karena aku

play13:26

udah update jadi jawaban nya pakai

play13:29

latest version aja yaitu

play13:31

260 dan sebenarnya sebenarnya bisa

play13:35

berapa data sih yang kita

play13:37

dapatin jujur enggak tahu dan Harusnya

play13:40

bisa banyak asalkan kalian sabar nunggu

play13:43

aja Soalnya aku nunggu Aku gak tahu

play13:46

nunggu berapa lama itu itu aja udah

play13:48

dapetin

play13:49

2.700 dan ini 2700 ini e aku stop manual

play13:55

gitu jadi kan ini habis kalau kita Run

play13:58

gitu kan B ada ion stop gitu ya itu aku

play14:01

stop sendiri

play14:03

soalnya takut ke keblokir juga nih

play14:06

twitterku karena Scroll terlalu lama

play14:08

gitu ya jadi disesuaikan kebutuhan

play14:11

kalian aja harusnya 10.000 gitu harusnya

play14:14

bisa sih Pokoknya ditunggu aja ya gitu

play14:19

ya harusnya main objektif 3 video ini

play14:22

udah selesai karenaudah menjawab topik

play14:24

ini mungk berikutnya Kalau tem-an ada

play14:27

pertanyaan yangin semoga aku bisa Jawab

play14:29

please drop the question di komentar di

play14:33

video

play14:34

ini tapi mungkin ini aku short answer

play14:37

aja TF itu apa tadi aku udah jelasin dia

play14:41

tool buat collnya cara kerjanya mungkin

play14:44

nanti aku bakal bikin di video sebelah

play14:47

kalau ada yang

play14:48

butuh nah ini cukup menarik bisa buat

play14:51

social network analisis atau enggak

play14:53

jawabannya bisa di video terakhir eh di

play14:57

versi yang terakhir ini bisa nanti bakal

play15:00

aku bahas kalau mau cral data pakai

play15:03

bahasa Indonesia aja bisa enggak

play15:05

keyword-nya di langang ti2 ID Jadi kalau

play15:08

kalian mau cari tentang Presiden gitu ya

play15:12

berbahasa Indonesia Tinggal Kasih Lang

play15:14

ID

play15:16

gitu kalau dapatin data dari rentang

play15:18

waktu atau bisa atau enggak bisa

play15:20

presiden misalnya Jokowi gitu ya tentang

play15:23

Pak

play15:24

Jokowi terus pakai

play15:26

Sin tahun berapa bulan apa tanggal

play15:30

berapa until Kapan 2024

play15:36

01 bulan 1 tanggal 30 misalkan gitu

play15:40

kalau dapatin data komentar bisa atau

play15:42

enggak bisa cek videoku yang

play15:46

sebelumnya Harusnya sama tinggal kalian

play15:48

Ganti versi ke 260 aja guu dapatin twet

play15:52

username Tertentu bisa atau enggak bisa

play15:54

Misalkan

play15:55

ee bahas tentang laut

play16:00

atau virus lah atau covid covid tapi

play16:04

dari dari Twitternya Pak Jokowi gitu nah

play16:08

keyword-nya jadi kayak gini jadi ditaruh

play16:11

di bagian search keyword

play16:13

nih

play16:17

gitu Nah kayak gini nih jadi ini bakal

play16:20

kita bakal nyari tweet yang ada

play16:22

kata-kata covidnya yang diweet sama Pak

play16:25

Jokowi

play16:27

gitu

play16:30

kalau dapatin data dengan minimum Lik

play16:32

atau reply gimana senya kalau

play16:34

teman-teman cek di blogku

play16:37

ya kalau mau Lebih detail

play16:41

lagi di mana ya Nah di sini nih yang

play16:46

blog yang data Twitter dengan python dan

play16:49

SN sebennya ini udah lama sih

play16:54

tapi

play16:57

akuwiteryudah lengkap banget kalau

play17:00

kalian mau cari tweet dari pengguna

play17:03

tertentu ini bisa pakai ini

play17:06

Hash terus nah yang menyebutkan pengguna

play17:11

tertentu terus pakai or Nah tadi yang

play17:14

since until ada di sini juga dipost di

play17:18

kota mana Within berapa kilomer ini bisa

play17:22

juga terus pakai

play17:25

net yang mengandung frasa tepat ini

play17:28

pakai petik

play17:29

minimum repes minimum likes ada juga nih

play17:33

jadi jawabannya ini ya minimum repes

play17:36

titik2 10 berarti minimal yang reply 10

play17:39

yang minimal like 100 Berarti kayak gini

play17:42

terus minimum retweets kayak gini gitu

play17:46

ya teman-teman bisa cek di blokku yang

play17:48

itu Cara preprocessing datanya kayak

play17:52

gimana Next Video

play17:54

aja ya gitu ya teman-teman semoga

play17:57

menjawab

play18:02

Twitter thank you

Rate This
★
★
★
★
★

5.0 / 5 (0 votes)

Ähnliche Tags
Twitter APIData crawlingTweet HarvestResearch toolsRate limitsCSV exportGoogle ColabPython scriptsSocial media analysisKeyword search
Benötigen Sie eine Zusammenfassung auf Englisch?