Add sphinx formatting to hyperlink methods (#898)

This commit is contained in:
Jeff Widman
2017-03-03 10:14:54 -08:00
committed by Dana Powers
parent b1f22b882a
commit d9283c1453
5 changed files with 47 additions and 45 deletions

View File

@@ -91,11 +91,11 @@ class SimpleClient(object):
Returns the leader for a partition or None if the partition exists
but has no leader.
UnknownTopicOrPartitionError will be raised if the topic or partition
is not part of the metadata.
LeaderNotAvailableError is raised if server has metadata, but there is
no current leader
Raises:
UnknownTopicOrPartitionError: If the topic or partition is not part
of the metadata.
LeaderNotAvailableError: If the server has metadata, but there is no
current leader.
"""
key = TopicPartition(topic, partition)
@@ -434,8 +434,8 @@ class SimpleClient(object):
Create an inactive copy of the client object, suitable for passing
to a separate thread.
Note that the copied connections are not initialized, so reinit() must
be called on the returned copy.
Note that the copied connections are not initialized, so :meth:`.reinit`
must be called on the returned copy.
"""
_conns = self._conns
self._conns = {}

View File

@@ -35,7 +35,7 @@ class KafkaConsumer(six.Iterator):
Arguments:
*topics (str): optional list of topics to subscribe to. If not set,
call subscribe() or assign() before consuming records.
call :meth:`.subscribe` or :meth:`.assign` before consuming records.
Keyword Arguments:
bootstrap_servers: 'host[:port]' string (or list of 'host[:port]'
@@ -119,7 +119,7 @@ class KafkaConsumer(six.Iterator):
session_timeout_ms (int): The timeout used to detect failures when
using Kafka's group management facilities. Default: 30000
max_poll_records (int): The maximum number of records returned in a
single call to poll(). Default: 500
single call to :meth:`.poll`. Default: 500
receive_buffer_bytes (int): The size of the TCP receive buffer
(SO_RCVBUF) to use when reading data. Default: None (relies on
system defaults). The java client defaults to 32768.
@@ -327,11 +327,11 @@ class KafkaConsumer(six.Iterator):
partitions (list of TopicPartition): Assignment for this instance.
Raises:
IllegalStateError: If consumer has already called subscribe()
IllegalStateError: If consumer has already called :meth:`.subscribe`.
Warning:
It is not possible to use both manual partition assignment with
assign() and group assignment with subscribe().
:meth:`.assign` and group assignment with :meth:`.subscribe`.
Note:
This interface does not support incremental assignment and will
@@ -349,12 +349,12 @@ class KafkaConsumer(six.Iterator):
def assignment(self):
"""Get the TopicPartitions currently assigned to this consumer.
If partitions were directly assigned using assign(), then this will
simply return the same partitions that were previously assigned.
If topics were subscribed using subscribe(), then this will give the
set of topic partitions currently assigned to the consumer (which may
be None if the assignment hasn't happened yet, or if the partitions are
in the process of being reassigned).
If partitions were directly assigned using :meth:`.assign`, then this
will simply return the same partitions that were previously assigned.
If topics were subscribed using :meth:`.subscribe`, then this will give
the set of topic partitions currently assigned to the consumer (which
may be None if the assignment hasn't happened yet, or if the partitions
are in the process of being reassigned).
Returns:
set: {TopicPartition, ...}
@@ -518,7 +518,7 @@ class KafkaConsumer(six.Iterator):
with any records that are available currently in the buffer,
else returns empty. Must not be negative. Default: 0
max_records (int, optional): The maximum number of records returned
in a single call to :meth:`poll`. Default: Inherit value from
in a single call to :meth:`.poll`. Default: Inherit value from
max_poll_records.
Returns:
@@ -630,10 +630,10 @@ class KafkaConsumer(six.Iterator):
def pause(self, *partitions):
"""Suspend fetching from the requested partitions.
Future calls to poll() will not return any records from these partitions
until they have been resumed using resume(). Note that this method does
not affect partition subscription. In particular, it does not cause a
group rebalance when automatic assignment is used.
Future calls to :meth:`.poll` will not return any records from these
partitions until they have been resumed using :meth:`.resume`. Note that
this method does not affect partition subscription. In particular, it
does not cause a group rebalance when automatic assignment is used.
Arguments:
*partitions (TopicPartition): Partitions to pause.
@@ -645,7 +645,7 @@ class KafkaConsumer(six.Iterator):
self._subscription.pause(partition)
def paused(self):
"""Get the partitions that were previously paused by a call to pause().
"""Get the partitions that were previously paused using :meth:`.pause`.
Returns:
set: {partition (TopicPartition), ...}
@@ -668,10 +668,10 @@ class KafkaConsumer(six.Iterator):
"""Manually specify the fetch offset for a TopicPartition.
Overrides the fetch offsets that the consumer will use on the next
poll(). If this API is invoked for the same partition more than once,
the latest offset will be used on the next poll(). Note that you may
lose data if this API is arbitrarily used in the middle of consumption,
to reset the fetch offsets.
:meth:`.poll`. If this API is invoked for the same partition more than
once, the latest offset will be used on the next :meth:`.poll`. Note
that you may lose data if this API is arbitrarily used in the middle of
consumption, to reset the fetch offsets.
Arguments:
partition (TopicPartition): Partition for seek operation
@@ -743,7 +743,7 @@ class KafkaConsumer(six.Iterator):
Topic subscriptions are not incremental: this list will replace the
current assignment (if there is one).
This method is incompatible with assign().
This method is incompatible with :meth:`.assign`.
Arguments:
topics (list): List of topics for subscription.
@@ -772,7 +772,7 @@ class KafkaConsumer(six.Iterator):
through this interface are from topics subscribed in this call.
Raises:
IllegalStateError: If called after previously calling assign().
IllegalStateError: If called after previously calling :meth:`.assign`.
AssertionError: If neither topics or pattern is provided.
TypeError: If listener is not a ConsumerRebalanceListener.
"""

View File

@@ -43,10 +43,10 @@ class BaseCoordinator(object):
leader and begins processing.
To leverage this protocol, an implementation must define the format of
metadata provided by each member for group registration in group_protocols()
and the format of the state assignment provided by the leader in
_perform_assignment() and which becomes available to members in
_on_join_complete().
metadata provided by each member for group registration in
:meth:`.group_protocols` and the format of the state assignment provided by
the leader in :meth:`._perform_assignment` and which becomes available to
members in :meth:`._on_join_complete`.
"""
DEFAULT_CONFIG = {
@@ -277,7 +277,7 @@ class BaseCoordinator(object):
"""Join the group and return the assignment for the next generation.
This function handles both JoinGroup and SyncGroup, delegating to
_perform_assignment() if elected leader by the coordinator.
:meth:`._perform_assignment` if elected leader by the coordinator.
Returns:
Future: resolves to the encoded-bytes assignment returned from the

View File

@@ -56,7 +56,8 @@ def _send_upstream(queue, client, codec, batch_time, batch_size,
Messages placed on the queue should be tuples that conform to this format:
((topic, partition), message, key)
Currently does not mark messages with task_done. Do not attempt to join()!
Currently does not mark messages with task_done. Do not attempt to
:meth:`join`!
Arguments:
queue (threading.Queue): the queue from which to get messages
@@ -227,7 +228,8 @@ class Producer(object):
Arguments:
client (kafka.SimpleClient): instance to use for broker
communications. If async=True, the background thread will use
client.copy(), which is expected to return a thread-safe object.
:meth:`client.copy`, which is expected to return a thread-safe
object.
codec (kafka.protocol.ALL_CODECS): compression codec to use.
req_acks (int, optional): A value indicating the acknowledgements that
the server must receive before responding to the request,
@@ -263,7 +265,7 @@ class Producer(object):
will not allow you to identify the specific message that failed,
but it will allow you to match failures with retries.
async_stop_timeout (int or float, optional): seconds to continue
attempting to send queued messages after producer.stop(),
attempting to send queued messages after :meth:`producer.stop`,
defaults to 30.
Deprecated Arguments:

View File

@@ -35,9 +35,9 @@ class KafkaProducer(object):
thread that is responsible for turning these records into requests and
transmitting them to the cluster.
The send() method is asynchronous. When called it adds the record to a
buffer of pending record sends and immediately returns. This allows the
producer to batch together individual records for efficiency.
:meth:`.send` is asynchronous. When called it adds the record to a buffer of
pending record sends and immediately returns. This allows the producer to
batch together individual records for efficiency.
The 'acks' config controls the criteria under which requests are considered
complete. The "all" setting will result in blocking on the full commit of
@@ -167,9 +167,9 @@ class KafkaProducer(object):
will block up to max_block_ms, raising an exception on timeout.
In the current implementation, this setting is an approximation.
Default: 33554432 (32MB)
max_block_ms (int): Number of milliseconds to block during send() and
partitions_for(). These methods can be blocked either because the
buffer is full or metadata unavailable. Blocking in the
max_block_ms (int): Number of milliseconds to block during :meth:`.send`
and :meth:`.partitions_for`. These methods can be blocked either
because the buffer is full or metadata unavailable. Blocking in the
user-supplied serializers or partitioner will not be counted against
this timeout. Default: 60000.
max_request_size (int): The maximum size of a request. This is also
@@ -537,8 +537,8 @@ class KafkaProducer(object):
Invoking this method makes all buffered records immediately available
to send (even if linger_ms is greater than 0) and blocks on the
completion of the requests associated with these records. The
post-condition of flush() is that any previously sent record will have
completed (e.g. Future.is_done() == True). A request is considered
post-condition of :meth:`.flush` is that any previously sent record will
have completed (e.g. Future.is_done() == True). A request is considered
completed when either it is successfully acknowledged according to the
'acks' configuration for the producer, or it results in an error.