Skip to content
Closed
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
2 changes: 1 addition & 1 deletion python/docs/pyspark.streaming.rst
Original file line number Diff line number Diff line change
@@ -1,5 +1,5 @@
pyspark.streaming module
==================
========================

Module contents
---------------
Expand Down
27 changes: 1 addition & 26 deletions python/pyspark/mllib/__init__.py
Original file line number Diff line number Diff line change
Expand Up @@ -32,29 +32,4 @@
import rand as random
random.__name__ = 'random'
random.RandomRDDs.__module__ = __name__ + '.random'


class RandomModuleHook(object):
"""
Hook to import pyspark.mllib.random
"""
fullname = __name__ + '.random'

def find_module(self, name, path=None):
# skip all other modules
if not name.startswith(self.fullname):
return
return self

def load_module(self, name):
if name == self.fullname:
return random

cname = name.rsplit('.', 1)[-1]
try:
return getattr(random, cname)
except AttributeError:
raise ImportError


sys.meta_path.append(RandomModuleHook())
sys.modules[__name__ + '.random'] = random
6 changes: 3 additions & 3 deletions python/pyspark/mllib/feature.py
Original file line number Diff line number Diff line change
Expand Up @@ -53,10 +53,10 @@ class Normalizer(VectorTransformer):
"""
:: Experimental ::

Normalizes samples individually to unit L\ :sup:`p`\ norm
Normalizes samples individually to unit L\ :sup:`p`\ norm

For any 1 <= `p` <= float('inf'), normalizes samples using
sum(abs(vector). :sup:`p`) :sup:`(1/p)` as norm.
For any 1 <= `p` < float('inf'), normalizes samples using
sum(abs(vector) :sup:`p`) :sup:`(1/p)` as norm.

For `p` = float('inf'), max(abs(vector)) will be used as norm for normalization.

Expand Down