Golgappa.net | Golgappa.org | BagIndia.net | BodyIndia.Com | CabIndia.net | CarsBikes.net | CarsBikes.org | CashIndia.net | ConsumerIndia.net | CookingIndia.net | DataIndia.net | DealIndia.net | EmailIndia.net | FirstTablet.com | FirstTourist.com | ForsaleIndia.net | IndiaBody.Com | IndiaCab.net | IndiaCash.net | IndiaModel.net | KidForum.net | OfficeIndia.net | PaysIndia.com | RestaurantIndia.net | RestaurantsIndia.net | SaleForum.net | SellForum.net | SoldIndia.com | StarIndia.net | TomatoCab.com | TomatoCabs.com | TownIndia.com
Interested to Buy Any Domain ? << Click Here >> for more details...

How would you determine the quantity of parcels while making a RDD? What are the capacities?

Answer Posted / Ms Varsha

In PySpark, the number of parcels in a Resilient Distributed Dataset (RDD) can be determined by using actions like count(). The capacity of each parcel can be defined during the creation of RDD, for example, when reading data from a file or a database. Here is an example of creating an RDD with predefined capacities:

rdd = sc.textFile("data.txt", 4)

In this example, the textFile function takes two arguments - the path to the data and the number of splits (capacities).

Is This Answer Correct ?    0 Yes 0 No



Post New Answer       View All Answers


Please Help Members By Posting Answers For Below Questions

How might you associate Hive to Spark SQL?

8