Spark Standalone Architecture

Ғылым және технология

#spark #bigdata #apachespark #hadoop #nosql #sparkwordcount #sparkarchitecture
Video Playlist
-----------------------
Hadoop in Tamil - bit.ly/32k6mBD
Hadoop in English - bit.ly/32jle3t
Spark in Tamil - bit.ly/2ZzWAJN
Spark in English - bit.ly/3mmc0eu
Batch vs Stream processing Tamil - • Data - Batch processi...
Batch vs Stream processing English - • Data - Batch processi...
NOSQL in English - bit.ly/2XtU07B
NOSQL in Tamil - bit.ly/2XVLLjP
Scala in Tamil : goo.gl/VfAp6d
Scala in English: goo.gl/7l2USl
Email : atozknowledge.com@gmail.com
LinkedIn : / sbgowtham
Instagram : / bigdata.in
KZread channel link
/ atozknowledgevideos
Website
atozknowledge.com/
Technology in Tamil & English
#bigdata #hadoop #spark #apachehadoop #whatisbigdata #bigdataintroduction #bigdataonline #bigdataintamil #bigdatatamil #hadoop #hadoopframework #hive #hbase #sqoop #mapreduce #hdfs #hadoopecosystem

Пікірлер: 23

  • @rohitbhagwat3031
    @rohitbhagwat30313 жыл бұрын

    Excellent!!!! Hats off to you teaching. Don't stop teaching. Each and every point got cleared. Keep it up.

  • @ittzzmalind9101
    @ittzzmalind91016 күн бұрын

    Bro I have a question. In this video data was distributed to two nodes even before executors are created ( before program logic starts ryt) But in my program if I read data from another system (like s3) during program execution only data will be loaded to cluster. So program execution ( creation of executor) should start first ryt before distributing data?

  • @Devarati
    @Devarati2 жыл бұрын

    Fantastic teaching with amazing clarity, point by point explanation. Thank you.

  • @PavanKumar-vi7hd
    @PavanKumar-vi7hd11 ай бұрын

    Your way of teaching is excellent Gowtham

  • @channuangadi8506
    @channuangadi8506 Жыл бұрын

    This is the best explanation i ever saw Very indepth explanation

  • @gsekhar4367
    @gsekhar43672 жыл бұрын

    It's crystal clear i like ur way of teaching

  • @sonip472
    @sonip472 Жыл бұрын

    Excellent video, very well explained.

  • @parthlodaria5656
    @parthlodaria56563 жыл бұрын

    Great Explanation!!

  • @bommanasravan8279
    @bommanasravan827921 күн бұрын

    Excellent

  • @RohitSaini52945
    @RohitSaini529452 жыл бұрын

    Thank you !

  • @raviy10
    @raviy102 жыл бұрын

    Thank you !!!

  • @sowjanyarbhat4715
    @sowjanyarbhat4715 Жыл бұрын

    Is it fine to use Spark Standalone for a POC project , or should I use with Hadoop itself? The requirement is basically to migrate Django Cron jobs what we run on Celery(with RabbitMQ) to Spark .

  • @bharathmamidisetti3896
    @bharathmamidisetti389614 күн бұрын

    Thalaiva❤❤❤❤

  • @manikandankbk7633
    @manikandankbk76333 жыл бұрын

    Bro, when I run in standalone mode with 1 master and 2 workers (in laptop, I have made host as master and first worker... (and) 2nd worker in VM)... Only when I put the same input file in both the worker node in gives me correct output. For eg. I need to put the same 1gb input file in both workers. If I partition them into 512mb (I did it manually, but the location of file in both worker is same) , first half of the file in one worker and other half in second worker, it gave incorrect results. Why is it so?

  • @atulsingh9202

    @atulsingh9202

    9 ай бұрын

    When manually partitioning data in a Spark standalone cluster, ensure that each partition resides on a separate worker node and has unique file paths to avoid data duplication and incorrect results

  • @gsekhar4367
    @gsekhar43672 жыл бұрын

    Brother here master and worker are deamons or physical servers?

  • @dhananjayreddy9998
    @dhananjayreddy99982 жыл бұрын

    In Standalone mode, if there is no requirement of splitting the file into multi nodes what is the purpose of having worker nodes(multi nodes).

  • @atulsingh9202

    @atulsingh9202

    9 ай бұрын

    In Spark's standalone cluster mode, worker nodes provide resource isolation, fault tolerance, scalability, parallelism, and efficient resource management, even if data splitting is not required.

  • @parammani4717
    @parammani47172 жыл бұрын

    Hi, ji can you upload spark architecture with yarn

  • @rajashekarm6162
    @rajashekarm6162 Жыл бұрын

    Fantastic brother, but in real-time most entities uses Spark with Yarn Deployment mode, these many months you still didn't upload that video.!!!

  • @dataengineeringvideos

    @dataengineeringvideos

    Жыл бұрын

    Hi bro Thanks Please find the video for yarn deployment in spark kzread.info/dash/bejne/ZZdqlI9_ZsXbZc4.html

  • @SANGEETHAD-mb2hj
    @SANGEETHAD-mb2hj2 жыл бұрын

    In general How will u read data line by line in python

  • @gsekhar4367

    @gsekhar4367

    2 жыл бұрын

    Really its crystal clear explantion i like ur videos way of explantion tq

Келесі