spark.memory.fraction expresses the size of M as a fraction of the (JVM heap space - 300MiB) (default 0.6). The rest of the space (40%) is reserved for user data structures, internal metadata in Spark, and safeguarding against OOM errors in the case of sparse and unusually large records.
I really appreciate your time and efforts in making quality videos. Please explain us how these different memory allocations cause problems or exceptions. How to solve these exceptions or problems. A screen shot of the possible issues and code/configuration changes to solve the issue will be really helpful and we would be really greatful if you could provide these details as well. Once again I appreciate your work and efforts
Hi Sir,the doubts fog clearing from mind after watching your spark videos,kindly make one session on real-time project from requirement to deployment it will very helpful ,Thank you.
Hi, Your videos are giving a good real time knowledge on spark and i thank you for that.. Could you please make a video on how to submit spark code(Pyspark) using shell script. also how to submit a spark job using shell script if both can be done differently. Thanks in advance
how do we find out if any executor is overallocated memory with --executor-memory but actually the job needs very less memory than provided executor memory parameter . Does this cause spark executer to reserve this memory and not being useful for other executors ?
If execution memory can evict blocks of data from storage memory, what happens to those evicted blocks if they are to be consumed again Will they be computed again and stored again
Hello sir, I have some questions if you could answer in free time when i read spark.read.csv(and provide inferSchema=True) Does it take all rows to guess the datatype of a column what is sampleRatio option in spark.read.csv ? is it related to infershema can i tell spark to use all rows while infering the schema for a column
Hi Harjeet, I am experienced professional who need some help in understanding current market conditions for future planning purposes (a kind of short call for mentoring). Please let me know if you can provide 10-15 min time.