Home
Trees
Indices
Help
Spark 1.0.2 Python API Docs
[
frames
] |
no frames
]
Identifier Index
[
A
B
C
D
E
F
G
H
I
J
K
L
M
N
O
P
Q
R
S
T
U
V
W
X Y
Z
_
]
A
Accumulator
(in
pyspark.accumulators
)
add()
(in
Accumulator
)
addInPlace()
(in
AddingAccumulatorParam
)
accumulator()
(in
SparkContext
)
addFile()
(in
SparkContext
)
addPyFile()
(in
SparkContext
)
AccumulatorParam
(in
pyspark.accumulators
)
AddingAccumulatorParam
(in
pyspark.accumulators
)
aggregate()
(in
RDD
)
accumulators
(in
pyspark
)
addInPlace()
(in
AccumulatorParam
)
ALS
(in
pyspark.mllib.recommendation
)
B
broadcast
(in
pyspark
)
Broadcast
(in
pyspark.broadcast
)
broadcast()
(in
SparkContext
)
C
cache()
(in
RDD
)
clusterCenters()
(in
KMeansModel
)
contains()
(in
SparkConf
)
cache()
(in
SchemaRDD
)
clustering
(in
pyspark.mllib
)
context
(in
pyspark
)
cacheTable()
(in
SQLContext
)
coalesce()
(in
RDD
)
context()
(in
RDD
)
cancelAllJobs()
(in
SparkContext
)
coalesce()
(in
SchemaRDD
)
copy()
(in
StatCounter
)
cancelJobGroup()
(in
SparkContext
)
cogroup()
(in
RDD
)
count()
(in
RDD
)
cartesian()
(in
RDD
)
collect()
(in
RDD
)
count()
(in
SchemaRDD
)
checkpoint()
(in
RDD
)
collectAsMap()
(in
RDD
)
count()
(in
StatCounter
)
checkpoint()
(in
SchemaRDD
)
combineByKey()
(in
RDD
)
countByKey()
(in
RDD
)
classification
(in
pyspark.mllib
)
COMPLEX_ACCUMULATOR_PARAM
(in
pyspark.accumulators
)
countByValue()
(in
RDD
)
clearFiles()
(in
SparkContext
)
conf
(in
pyspark
)
D
defaultMinPartitions()
(in
SparkContext
)
DISK_ONLY_2
(in
StorageLevel
)
dumps
(in
MarshalSerializer
)
defaultParallelism()
(in
SparkContext
)
distinct()
(in
RDD
)
dumps()
(in
PickleSerializer
)
dense()
(in
Vectors
)
distinct()
(in
SchemaRDD
)
DISK_ONLY
(in
StorageLevel
)
dot()
(in
SparseVector
)
F
files
(in
pyspark
)
flatMapValues()
(in
RDD
)
foreach()
(in
RDD
)
filter()
(in
RDD
)
FLOAT_ACCUMULATOR_PARAM
(in
pyspark.accumulators
)
foreachPartition()
(in
RDD
)
first()
(in
RDD
)
fold()
(in
RDD
)
flatMap()
(in
RDD
)
foldByKey()
(in
RDD
)
G
get()
(in
SparkConf
)
getCheckpointFile()
(in
SchemaRDD
)
glom()
(in
RDD
)
get()
(in
SparkFiles
)
getLocalProperty()
(in
SparkContext
)
groupBy()
(in
RDD
)
getAll()
(in
SparkConf
)
getRootDirectory()
(in
SparkFiles
)
groupByKey()
(in
RDD
)
getCheckpointFile()
(in
RDD
)
getStorageLevel()
(in
RDD
)
groupWith()
(in
RDD
)
H
HiveContext
(in
pyspark.sql
)
hiveql()
(in
HiveContext
)
hql()
(in
HiveContext
)
I
id()
(in
RDD
)
INT_ACCUMULATOR_PARAM
(in
pyspark.accumulators
)
intersection()
(in
SchemaRDD
)
inferSchema()
(in
SQLContext
)
intercept()
(in
LinearModel
)
isCheckpointed()
(in
RDD
)
insertInto()
(in
SchemaRDD
)
intersection()
(in
RDD
)
isCheckpointed()
(in
SchemaRDD
)
J
join()
(in
RDD
)
jsonFile()
(in
SQLContext
)
jsonRDD()
(in
SQLContext
)
K
keyBy()
(in
RDD
)
KMeans
(in
pyspark.mllib.clustering
)
keys()
(in
RDD
)
KMeansModel
(in
pyspark.mllib.clustering
)
L
LabeledPoint
(in
pyspark.mllib.regression
)
LinearModel
(in
pyspark.mllib.regression
)
loads
(in
MarshalSerializer
)
LassoModel
(in
pyspark.mllib.regression
)
LinearRegressionModel
(in
pyspark.mllib.regression
)
loads
(in
PickleSerializer
)
LassoWithSGD
(in
pyspark.mllib.regression
)
LinearRegressionModelBase
(in
pyspark.mllib.regression
)
LocalHiveContext
(in
pyspark.sql
)
leftOuterJoin()
(in
RDD
)
LinearRegressionWithSGD
(in
pyspark.mllib.regression
)
LogisticRegressionModel
(in
pyspark.mllib.classification
)
linalg
(in
pyspark.mllib
)
loadLibSVMFile()
(in
MLUtils
)
LogisticRegressionWithSGD
(in
pyspark.mllib.classification
)
M
map()
(in
RDD
)
mean()
(in
RDD
)
MEMORY_ONLY_SER_2
(in
StorageLevel
)
mapPartitions()
(in
RDD
)
mean()
(in
StatCounter
)
merge()
(in
StatCounter
)
mapPartitionsWithIndex()
(in
RDD
)
MEMORY_AND_DISK
(in
StorageLevel
)
mergeStats()
(in
StatCounter
)
mapPartitionsWithSplit()
(in
RDD
)
MEMORY_AND_DISK_2
(in
StorageLevel
)
min()
(in
RDD
)
mapValues()
(in
RDD
)
MEMORY_AND_DISK_SER
(in
StorageLevel
)
min()
(in
StatCounter
)
MarshalSerializer
(in
pyspark.serializers
)
MEMORY_AND_DISK_SER_2
(in
StorageLevel
)
mllib
(in
pyspark
)
MatrixFactorizationModel
(in
pyspark.mllib.recommendation
)
MEMORY_ONLY
(in
StorageLevel
)
MLUtils
(in
pyspark.mllib.util
)
max()
(in
RDD
)
MEMORY_ONLY_2
(in
StorageLevel
)
max()
(in
StatCounter
)
MEMORY_ONLY_SER
(in
StorageLevel
)
N
NaiveBayes
(in
pyspark.mllib.classification
)
NaiveBayesModel
(in
pyspark.mllib.classification
)
name()
(in
RDD
)
O
OFF_HEAP
(in
StorageLevel
)
P
parallelize()
(in
SparkContext
)
PickleSerializer
(in
pyspark.serializers
)
predict()
(in
MatrixFactorizationModel
)
parquetFile()
(in
SQLContext
)
pipe()
(in
RDD
)
predict()
(in
LinearRegressionModelBase
)
partitionBy()
(in
RDD
)
predict()
(in
LogisticRegressionModel
)
predictAll()
(in
MatrixFactorizationModel
)
persist()
(in
RDD
)
predict()
(in
NaiveBayesModel
)
printSchema()
(in
SchemaRDD
)
persist()
(in
SchemaRDD
)
predict()
(in
SVMModel
)
pyspark
pickleSer
(in
pyspark.accumulators
)
predict()
(in
KMeansModel
)
R
rdd
(in
pyspark
)
registerAsTable()
(in
SchemaRDD
)
ResultIterable
(in
pyspark.resultiterable
)
RDD
(in
pyspark.rdd
)
registerRDDAsTable()
(in
SQLContext
)
RidgeRegressionModel
(in
pyspark.mllib.regression
)
recommendation
(in
pyspark.mllib
)
regression
(in
pyspark.mllib
)
RidgeRegressionWithSGD
(in
pyspark.mllib.regression
)
reduce()
(in
RDD
)
repartition()
(in
RDD
)
rightOuterJoin()
(in
RDD
)
reduceByKey()
(in
RDD
)
repartition()
(in
SchemaRDD
)
Row
(in
pyspark.sql
)
reduceByKeyLocally()
(in
RDD
)
resultiterable
(in
pyspark
)
S
sample()
(in
RDD
)
setJobGroup()
(in
SparkContext
)
statcounter
(in
pyspark
)
sampleStdev()
(in
RDD
)
setLocalProperty()
(in
SparkContext
)
StatCounter
(in
pyspark.statcounter
)
sampleStdev()
(in
StatCounter
)
setMaster()
(in
SparkConf
)
stats()
(in
RDD
)
sampleVariance()
(in
RDD
)
setName()
(in
RDD
)
stdev()
(in
RDD
)
sampleVariance()
(in
StatCounter
)
setSparkHome()
(in
SparkConf
)
stdev()
(in
StatCounter
)
saveAsLibSVMFile()
(in
MLUtils
)
setSystemProperty()
(in
SparkContext
)
stop()
(in
SparkContext
)
saveAsParquetFile()
(in
SchemaRDD
)
sortByKey()
(in
RDD
)
storagelevel
(in
pyspark
)
saveAsTable()
(in
SchemaRDD
)
SparkConf
(in
pyspark.conf
)
StorageLevel
(in
pyspark.storagelevel
)
saveAsTextFile()
(in
RDD
)
SparkContext
(in
pyspark.context
)
subtract()
(in
RDD
)
SchemaRDD
(in
pyspark.sql
)
SparkFiles
(in
pyspark.files
)
subtract()
(in
SchemaRDD
)
schemaString()
(in
SchemaRDD
)
sparkUser()
(in
SparkContext
)
subtractByKey()
(in
RDD
)
serializers
(in
pyspark
)
sparse()
(in
Vectors
)
sum()
(in
RDD
)
set()
(in
SparkConf
)
SparseVector
(in
pyspark.mllib.linalg
)
sum()
(in
StatCounter
)
setAll()
(in
SparkConf
)
sql
(in
pyspark
)
SVMModel
(in
pyspark.mllib.classification
)
setAppName()
(in
SparkConf
)
sql()
(in
SQLContext
)
SVMWithSGD
(in
pyspark.mllib.classification
)
setCheckpointDir()
(in
SparkContext
)
SQLContext
(in
pyspark.sql
)
setExecutorEnv()
(in
SparkConf
)
squared_distance()
(in
SparseVector
)
T
table()
(in
SQLContext
)
toDebugString()
(in
SparkConf
)
train()
(in
KMeans
)
take()
(in
RDD
)
toDebugString()
(in
RDD
)
train()
(in
ALS
)
takeOrdered()
(in
RDD
)
top()
(in
RDD
)
train()
(in
LassoWithSGD
)
takeSample()
(in
RDD
)
train()
(in
LogisticRegressionWithSGD
)
train()
(in
LinearRegressionWithSGD
)
TestHiveContext
(in
pyspark.sql
)
train()
(in
NaiveBayes
)
train()
(in
RidgeRegressionWithSGD
)
textFile()
(in
SparkContext
)
train()
(in
SVMWithSGD
)
trainImplicit()
(in
ALS
)
U
uncacheTable()
(in
SQLContext
)
union()
(in
RDD
)
unpersist()
(in
SchemaRDD
)
union()
(in
SparkContext
)
unpersist()
(in
RDD
)
util
(in
pyspark.mllib
)
V
value()
(in
Accumulator
)
variance()
(in
RDD
)
Vectors
(in
pyspark.mllib.linalg
)
values()
(in
RDD
)
variance()
(in
StatCounter
)
W
weights()
(in
LinearModel
)
wholeTextFiles()
(in
SparkContext
)
Z
zero()
(in
AccumulatorParam
)
zero()
(in
AddingAccumulatorParam
)
zip()
(in
RDD
)
_
__add__()
(in
RDD
)
__init__()
(in
RDD
)
__str__()
(in
Accumulator
)
__del__()
(in
SparkContext
)
__init__()
(in
ResultIterable
)
__str__()
(in
SparseVector
)
__del__()
(in
MatrixFactorizationModel
)
__init__()
(in
Row
)
_accumulatorRegistry
(in
pyspark.accumulators
)
__eq__()
(in
SparseVector
)
__init__()
(in
SQLContext
)
_active_spark_context
(in
SparkContext
)
__iadd__()
(in
Accumulator
)
__init__()
(in
SchemaRDD
)
_broadcastRegistry
(in
pyspark.broadcast
)
__init__()
(in
Accumulator
)
__init__()
(in
StatCounter
)
_gateway
(in
SparkContext
)
__init__()
(in
AddingAccumulatorParam
)
__init__()
(in
StorageLevel
)
_is_running_on_worker
(in
SparkFiles
)
__init__()
(in
Broadcast
)
__iter__()
(in
ResultIterable
)
_jvm
(in
SparkContext
)
__init__()
(in
SparkConf
)
__len__()
(in
ResultIterable
)
_lock
(in
SparkContext
)
__init__()
(in
SparkContext
)
__ne__()
(in
SparseVector
)
_next_accum_id
(in
SparkContext
)
__init__()
(in
SparkFiles
)
__reduce__()
(in
Accumulator
)
_python_includes
(in
SparkContext
)
__init__()
(in
NaiveBayesModel
)
__reduce__()
(in
Broadcast
)
_root_directory
(in
SparkFiles
)
__init__()
(in
KMeansModel
)
__repr__()
(in
Accumulator
)
_sc
(in
SparkFiles
)
__init__()
(in
SparseVector
)
__repr__()
(in
SparseVector
)
_spark_stack_depth
(in
pyspark.rdd
)
__init__()
(in
MatrixFactorizationModel
)
__repr__()
(in
RDD
)
_writeToFile
(in
SparkContext
)
__init__()
(in
LabeledPoint
)
__repr__()
(in
StatCounter
)
__init__()
(in
LinearModel
)
__repr__()
(in
StorageLevel
)
Home
Trees
Indices
Help
Spark 1.0.2 Python API Docs
Generated by Epydoc 3.0.1 on Fri Jul 25 21:13:27 2014
http://epydoc.sourceforge.net