Hi, it is a public dataset only, I missed the source website url, my apologies. Please send us your email address to datamaking.training@gmail.com. We will send you the dataset. Thank you.
C:\Windows\System32>spark-shell Setting default log level to "WARN". To adjust logging level use sc.setLogLevel(newLevel). For SparkR, use setLogLevel(newLevel). 24/09/26 11:27:59 WARN NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable 24/09/26 11:27:59 ERROR Main: Failed to initialize Spark session. java.lang.UnsupportedOperationException: getSubject is supported only if a security manager is allowed
hi sir, i am getting this error, can you give me the solution? Failed to get schema version. Underlying cause: java.sql.SQLNonTransientConnectionException : Could not create connection to database server. Attempted reconnect 3 times. Giving up. SQL Error code: 0 Use --verbose for detailed stacktrace. *** schemaTool failed *** --verbose org.apache.hadoop.hive.metastore.HiveMetaException: Failed to get schema version. Caused by: java.sql.SQLNonTransientConnectionException: Could not create connection to database server. Attempted reconnect 3 times. Giving up. Caused by: com.mysql.cj.exceptions.CJException: Public Key Retrieval is not allowed Caused by: com.mysql.cj.exceptions.UnableToConnectException: Public Key Retrieval is not allowed *** schemaTool failed ***
Hi, please reach out to datamaking.training@gmail.com. Team will reach out to you soon. Thank you for showing interest in my video technical content. Appreciate it. Thank you 🙏
For those who don.t whan't the anaconda environment by default each time you open a terminal run this 'conda config --set auto_activate_base false', this will let you open terminals as usual and invoke conda only when desired
for all who gets error "Microsoft ODBC Driver 18 for SQL Server : Client unable to establish connection." after executing "/opt/mssql-tools/bin/sqlcmd...", just add "-C" at the end of the line.
Hi, Thanks for your wonderful content !, somehow actual video code is not in the downloaded resource. and save_to_mysql is not cleared could you please upload the actual content what you taught on this video, In the latest script of realtime_data_processing.py "cassandra part" is totally missing. please consider @DataMaking
@@Nalla-perumal Hi, thank you for showing interest. Unfortunately I don't have this source since I created very long back. I am planning to re-create in future. Thank you.
@@lucasgonzalezsonnenberg3204 thank you for showing on my technical content. I have another video but little older. FYI. ru-vid.com/video/%D0%B2%D0%B8%D0%B4%D0%B5%D0%BE-NY3IqbQSa-k.htmlsi=YP_weidRL7LbUeWM
@@siddharthat88 Hi, if you want to connect remote Spark cluster, you can use spark standalone cluster URL or resource manager URL based on the setup. Additionally some security property also need to provide. You can provide more details accordingly we can discuss here. Thank you 🙏
I am a new bee in the Data Engineering domain. I have idea on the most of the tech stacks of big data application but I was looking to connect all the dots. Your explanation was really very nice and smooth and the example also very relatable with current scenario world. Awesome content. Please keep uploading such good stuffs 🙏🙏
@@akindia8519 Hi, this is more on data engineering, but you can use PySpark technology as one of skill set for data analyst job when you want to analysis big data(large data set). Thank you 🙏
I'm getting the following error can somebody help hadoop@rahulparihar-Inspiron-N5010:~$ hive Hive Session ID = 661e5791-5e5d-497e-83cc-41167373ad0a Exception in thread "main" java.lang.ClassCastException: class jdk.internal.loader.ClassLoaders$AppClassLoader cannot be cast to class java.net.URLClassLoader (jdk.internal.loader.ClassLoaders$AppClassLoader and java.net.URLClassLoader are in module java.base of loader 'bootstrap') at org.apache.hadoop.hive.ql.session.SessionState.<init>(SessionState.java:413) at org.apache.hadoop.hive.ql.session.SessionState.<init>(SessionState.java:389) at org.apache.hadoop.hive.cli.CliSessionState.<init>(CliSessionState.java:60) at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:705) at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:683) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:566) at org.apache.hadoop.util.RunJar.run(RunJar.java:328) at org.apache.hadoop.util.RunJar.main(RunJar.java:241)
Hi, thank you for your feedback. I will try it out in future video series. At the same time, few things I want to mention 1. Real time we work in Linux environment only except few development tasks. 2. I don't have windows environment with me I respect your ask, i will try it out in future video series 🙏
Hi. Your video is very good. You clearly provided the steps to be easy understand. I need your text editor which give every step. Can you upload to the subscribe link.
Pin this For those of you who got an error after the --verbose part. here's the fix. First off one of the xml file he provided has 'Hive' as user and 'Datamaking' as password. Go back and change that to your username and password (your mysql user and pass) After that its likely still failing because of the permissions to retrieve key are denied. To fix that, 1. Connect to mysql cli in ur terminal: mysql -u your_username -p #enter that password 2. Once you're logged into the MySQL command-line client, execute the ALTER USER command to change the authentication plugin for the desired user. Here's the command: ALTER USER 'your_user'@'localhost' IDENTIFIED WITH mysql_native_password BY 'your_password'; 3. Thats it. Run the command again and after like 2 minutes of scrolling and generating shit, it will complete the beeline thing.
Hi, sorry for the inconveniences. Are you getting any issues or error, please share with me at indiacloudtv@gmail.com Meanwhile I will also try to check it.
It is also depends on what output mode you are using: spark.apache.org/docs/latest/structured-streaming-programming-guide.html#output-modes Complete mode is bound to break if you data is growing.