Related Article: www.geeksforgeeks.org/twitter-sentiment-analysis-using-python/
@shlokkumar62577 ай бұрын
siddharthan sir, congratuations for joining geeks for geeks.
@alexramos58711 ай бұрын
Great content from Siddhardhan.
@Sany19205 ай бұрын
you got it bro!
@Sagardeep_Das2 ай бұрын
Thank you so much gfg. I even applied neutral tweets in doing this sentiment analysis. I will try to do this for other social media datasets.
@codehub832611 ай бұрын
Pls make a video on career recommendation after secondary school using ML
@bharathikathiresan8 ай бұрын
Very Clearly explanation
@sunshineandrainbow54534 ай бұрын
Beautifully explained!!! Thanks a lot sir 🙏🏻🙏🏻🙏🏻 PS : How can we use it to make a project which will take input from the user
@janaraj2958 ай бұрын
Nice explanation, completed the project👍
@rasdhdg8 ай бұрын
Hi brother. Can you please provide me the code . I will pay you the amount you demand . 🙏🙏🙏🙏🙏🙏
@1anu_ra-dha.8 ай бұрын
Have you run this project??
@janaraj2958 ай бұрын
@@1anu_ra-dha. Yes
@janaraj2958 ай бұрын
@@1anu_ra-dha. Yes
@reema90886 ай бұрын
Have U run the project ?
@rachanikhilrnr11 ай бұрын
But what if I want to test on new data
@sumitsamanta7416Ай бұрын
u said abt overfitting and we can clearly see the model is overfitted so can we improve the accuracy and make the model little underfit ?
@rahulmahatha908611 ай бұрын
link to the notebook please....
@AbhimanyuKaushik-i5p3 ай бұрын
stemming would remove the comma as well as the apostrophe but we need to keep the apostrophe . so what should be the re for that
@ravindarmadishetty73611 ай бұрын
Nice explanation
@janmjaysingh27197 ай бұрын
How can we use our own tweet or comment to test the machine. As for now we are using index as median. Tell me if i can actually write my own sentence and check whether its negative or positive?😢
@ymmich21434 ай бұрын
I have the same doubt. Please let me know as well if you have found the answer.
@ChithraiSelvan-d8y2 ай бұрын
@@ymmich2143 I aslo having same doubts
@ChithraiSelvan-d8y2 ай бұрын
I
@DeionDz12 күн бұрын
Convert your tweet into a vector. using the process you showed and then follow the same predication pattern.
@crickethighlights71858 ай бұрын
Thanks for the video
@saicharanchepuri81454 ай бұрын
Just a suggestion it would have been simpler if you directly utilized the methods as they are from the libraries rather than storing each method in a different variable. It kind of makes the code look heavy and also confusing and difficult to follow. Thanks for the tutorial.
@prasadmahadik35143 ай бұрын
Hi Sir, Can we split data in X customer_Id with Text ? So we can get which customers said negative reviews ?
8 ай бұрын
I have implemented the above code it took some time to stem then I implemented another code with lemmatization it was fast. In general, Lemmas are more computationally expensive than Stemming, when I checked this code again I found a potential change - for each loop, it's calling StopWords instead you can define it once within some variable and call that variable inside the loop. After making these changes it took me 2 mins for stemming using the above code.
@soukarya_ghosh66128 ай бұрын
can u give the dataset link ?
8 ай бұрын
@@soukarya_ghosh6612 You can find it in video itself at 3:30
@thegeeks20026 ай бұрын
can you please tell me how's this part of code after changes you mentioned?
6 ай бұрын
@@thegeeks2002 Before it took around 1 hour I guess after these changes it took 2 mins, However I ended up using Lematization even it took 2mins. But this tutorial is really helpful for me to get started with this usecase and then explore more advanced concepts for it.
@shreyanshgupta15876 ай бұрын
do you have the code with you rn??
@saurabhkushwah53883 ай бұрын
Greate Exeplaination sir.
@syedasabeeh-g4j8 ай бұрын
very useful. but my code does not showing stemmed_content and it's taking only 2 to 3 mins to run instead of 55 mins. what can be the error sir
@sizzrizz60748 ай бұрын
You didn't add blank space between the quotation marks at the .join statement.
@tejiriamrasa32587 ай бұрын
@@sizzrizz6074 that isn't the cause of the problem.
@anuragpaul4538 күн бұрын
sir, there are other options to make this step faster. it took me only 5 mins to perform stemming. Here's the code: import pandas as pd import re from nltk.corpus import stopwords from nltk.stem.porter import PorterStemmer from concurrent.futures import ThreadPoolExecutor from tqdm import tqdm # Initialize stopwords stop_words = set(stopwords.words('english')) def stemming(content): try: port_stem = PorterStemmer() # Instantiate inside the function stemmed_content = re.sub('[^a-zA-Z]', ' ', content).lower() return ' '.join(port_stem.stem(word) for word in stemmed_content.split() if word not in stop_words) except Exception as e: print(f"Error processing content: {content}. Error: {e}") return "" # Return an empty string on error def process_data(df): with ThreadPoolExecutor() as executor: return list(tqdm(executor.map(stemming, df['text']), total=len(df))) # Process the DataFrame in chunks chunk_size = 50000 # Adjust based on your memory capacity num_chunks = len(twitter_data) // chunk_size + 1 stemmed_contents = [] for i in tqdm(range(num_chunks)): start = i * chunk_size end = min((i + 1) * chunk_size, len(twitter_data)) chunk = twitter_data.iloc[start:end] stemmed_chunk = process_data(chunk) stemmed_contents.extend(stemmed_chunk) # Add the stemmed content back to the DataFrame twitter_data['stemmed_content'] = stemmed_contents
@eedemallika5055 ай бұрын
Great Explanation
@gaganm663911 ай бұрын
How can we increase the accuracy of the model
@shaikhahsan1007 ай бұрын
try different algorithms, maybe some other algorithm works better and gives a better accuracy. Also, if the dataset is not vast like the one on the video try gathering atleast 1000-10000 of data and train again to check the accuracy.
@albingeorgekurian43967 ай бұрын
@@shaikhahsan100 have any other suggesstion
@Revathi-hb9rs4 ай бұрын
sir,please provide the complete end to end neural network style transfer projects with web application
@subratbhattarai35186 ай бұрын
Is there a way to convert the vectorized numbers to string again?
@GooSoft3 ай бұрын
Which library is used like tool?
@growingfire4 ай бұрын
Siddharthan Sir
@ritvikraj83846 ай бұрын
Can you send the PPT file which you showed in the beginning of the video?
@tarandeep68615 ай бұрын
Great eplanation thanks a lot
@Gharwali_Baatein10 ай бұрын
sir i have tried the same dataset with same code but the accuracy score is much different then yours ,its causing an overfitting condition ,how it can be improved
@saurabhojha283210 ай бұрын
Kitni aayi thi bhaii?
@saurabhojha283210 ай бұрын
Btao
@Gharwali_Baatein10 ай бұрын
55%@@saurabhojha2832
@sohailaborhamy78556 ай бұрын
The same problem,you solve it or not ?
@ssenthilnathan3 Жыл бұрын
Is this Siddharth? That guy who has a telegram channel named Machine Learning and also a youtube channel? If so! Great to see you buddy❤..
@FarrelHaryomukti2 ай бұрын
thanks.
@mdafroj98469 ай бұрын
Hi can help me i got error at the line 20 of code twitter data['stemmed_content']=Twitter_data['text'].apply(stemming)
@srijantiwari92329 ай бұрын
Solve hua?
@raunakkakkar12318 ай бұрын
it should be twitter_data not Twitter_data (no capital T)
@suyashsawant89287 ай бұрын
@@raunakkakkar1231 Hi can you help me too I had an error at the same line twitter_data['stemmed_content'] = twitter_data['text'].apply(stemming) and the error was NameError : name 'twitter_data' is not defined
@bindhuaddepalli5803 ай бұрын
@@suyashsawant8928 coz there is no file named twitter data with you
@elbechir3511 ай бұрын
Sır! If we don't have the target column in our Twitter dataset, what should we do?
@albingeorgekurian43967 ай бұрын
yes we do..check whether u have selected sentiment140 from kaggle
@jananeejananee278610 ай бұрын
siddharthan is here(as instructor)
@valac20857 ай бұрын
the way he saved model was not enough, you have to lay a data pipeline for new data and also save the vectorizer
@karishmarajput18714 ай бұрын
how to add kaggle.json Path in Jupyter notebook ............please reply?
@AbhimanyuKaushik-i5p3 ай бұрын
what is up and down sampling
@rishikeshnanaware9 ай бұрын
X has 179715 features, but LogisticRegression is expecting 460873 features as input. I am getting this error while executing predict function on test data
@archana24677 ай бұрын
How you solved this error
@rishikeshnanaware7 ай бұрын
@@archana2467 for training data use fit_transform function and for testing data use transform() function
@rishikeshnanaware7 ай бұрын
@@archana2467 use fit_transfer function for training data and transfer function for testing data instead of using fit_transfer for testing data
@rishikeshnanaware7 ай бұрын
@@archana2467 use fit_transform function for training data and transfer function for testing data instead of using fit_transform for testing data
@vildanbozkurt27626 ай бұрын
@@archana2467 ı had the same mistake. random_state should be equal to 2 or yo should check taht you have written the X_test equation correctly in the "converting the textual data to numerical data" section . I mean ; X_test = vectorizer.transform(X_test)
@AbhimanyuKaushik-i5p3 ай бұрын
what to do when null values are found in dataset
@manjukumari33135 ай бұрын
please share the collab notebook link
@sizzrizz60748 ай бұрын
Can anyone help my prog shows X_test not defined when running the pickle file in another window . What to do
@UchihaItachiKonohagakure5 ай бұрын
check that line of code where you split the dataset into training and testing sets. where the random state = 2 was mentioned. Also check whether you executed that snippet or not. that could also be an error.
@divyaagarwal356311 ай бұрын
At the time of fetching api from kaggle I am getting a key error: content length Pls help
@garima21587 ай бұрын
stemming process is taking too long time to execute, due to which i'm not able to apply modelling
@albingeorgekurian43967 ай бұрын
we need to optimize the stemming function. like in list comprehension we are using stopwords.words('english') instead of that create a variable outside the function and use that variable instead.. here I guess it will optimize the code than the previous version
@jkscout17 күн бұрын
I didn't know 1.6 million was actually 16 million.
@shivanshmaurya85202 ай бұрын
Twitter API is not free, what to do now ??
@sohailaborhamy78556 ай бұрын
The output of counts_value not the same your output, why this ?
@rahulmahatha908611 ай бұрын
can you provide the source code please
@navneet-dubey75Ай бұрын
Here from Seasons of AI 2024👋🏼
@SWAMINATHANS-id4so5 ай бұрын
sir how to get colab lin
@yhd080810 ай бұрын
1:08:00
@charchitsharma3535Ай бұрын
can anybody tell where is the collab file link
@SWAMINATHANS-id4so5 ай бұрын
sir how to get this project coalab sheet. you told me that link in description.i searched it,i did not get
@deepikaravi92924 ай бұрын
Go and search for google colab in google and then click on new notebook .so u will get it
@mganesh84739 ай бұрын
sir why we are not perfoming lemmatization
@javeedma27648 ай бұрын
based on requirement.lemmatization is some what slow when compared to stemming
@yhd080810 ай бұрын
51:27
@letseducate1233 ай бұрын
Where is the code??
@SWAMINATHANS-id4so5 ай бұрын
how to get colab link
@T1s_kashyap3 ай бұрын
ho siddharthan sir 😁😁😁😁😁😁😁😁
@154yashraj34 ай бұрын
snaives bayes multinomial model has a higher accuracy
@monemtajbid3363Ай бұрын
can you plz give me all souce code of this project
@AbhishekKumar-wc4nx7 ай бұрын
can anyone tell me where we can find the all code in this is project
@GeeksforGeeksVideos7 ай бұрын
You can find everything in this article : www.geeksforgeeks.org/twitter-sentiment-analysis-using-python/
@ak47gamer8109 ай бұрын
28:14
@boleto74672 ай бұрын
Can you share the codes
@yuktasharma7921Ай бұрын
Can u please provide the source code?
@BTS__Army182 ай бұрын
Where is the code
@deepikat45208 ай бұрын
my training data accuracy is 99% and test data accuracy is 50 my model is overfitted how to avoid it
@BrainyiFY_o08 ай бұрын
L1 and L2 regularization are techniques used to prevent overfitting by adding a penalty term to the loss function during training
@rasdhdg8 ай бұрын
Hi brother. Can you please provide me the code . I will pay you the amount you demand . 🙏🙏🙏🙏🙏🙏
@Vkg_Universe8 ай бұрын
I also face the same problem
@deepikaravi92924 ай бұрын
50.22
@1anu_ra-dha.8 ай бұрын
How to deploy this model on web?? Can anybody please help....?? 🙏🙏
@1anu_ra-dha.8 ай бұрын
@cll1670 thanks for help
@albingeorgekurian43967 ай бұрын
what about creating an interface This is just a suggestion if u have any other ideas please share them.
@chahatverma40857 ай бұрын
22.25 code shows error. can anyone help?
@albingeorgekurian43967 ай бұрын
can u elaborate more
@Alex-oz5kn Жыл бұрын
It's ML not YemYel
@jubayer16408 ай бұрын
😄
@OfficeUse-oo4kx8 ай бұрын
What the point you considered 😅
@officialteflon59717 ай бұрын
😂😂😂
@Zainab-en9fd8 ай бұрын
Can anyone tell me is twitter api free
@thedailymotivational7 ай бұрын
No
@ChandraShekarYadav-qp8pz3 ай бұрын
develop brain tumor detection project using cnn
@albingeorgekurian43967 ай бұрын
hello friends i have completed the more than half tutorial still need to go.. in the tutorial I have seen that the stemming is taking to much time (almost 50m) so I just optimized it little bit.... code: [ ] pattern =re.compile('[^a-zA-Z]') [ ] english_stopwords = stopwords.words('english') [ ] port_stemmer = PorterStemmer() [ ] def stemming(content): stemmed_content = re.sub(pattern,' ',content) stemmed_content = stemmed_content.lower() stemmed_content = stemmed_content.split() stemmed_content = [port_stemmer.stem(word) for word in stemmed_content if not word in english_stopwords] stemmed_content = ' '.join(stemmed_content) return stemmed_content [ ] twitter_data['stemmed_content'] = twitter_data['text'].apply(stemming) mine just completed in 6m
@infinity89826 ай бұрын
I can't thank you enough 🛐finally I can continue the video😭
@amogu_076 ай бұрын
@@infinity8982did u complete it bro is it working??
@vanshikajain30143 ай бұрын
can u provide the link of source code or share it with me
@yuktasharma7921Ай бұрын
Hey can u please provide the whole source code if u have done
@ashwinipatil57335 ай бұрын
hello I am making web app on it using flask but i am facing this error(tweepy.errors.Unauthorized: 401 Unauthorized 89 - Invalid or expired token.) because i have not basic account so can you help me to solve error .please reply