diff --git a/.sync_state b/.sync_state index bffd7ba..76e5102 100644 --- a/.sync_state +++ b/.sync_state @@ -1,4 +1,4 @@ { - "last_synced_sha": "110dad99af8ae1062af4fa99fc7559104d793680", - "last_sync_time": "2025-11-25T05:55:10.370139" + "last_synced_sha": "fb89303330ff94edcbd16cf47d253d7242225ba6", + "last_sync_time": "2025-11-25T06:15:14.524561" } \ No newline at end of file diff --git a/docs/api/apifuture.md b/docs/api/apifuture.md index f6059b3..d70e8c3 100644 --- a/docs/api/apifuture.md +++ b/docs/api/apifuture.md @@ -1,5 +1,3 @@ -# `tinker.lib.public_interfaces.api_future` - API Future classes for handling async operations with retry logic. ## `APIFuture` Objects @@ -41,7 +39,6 @@ result = future.result() #### `result_async` ```python -@abstractmethod async def result_async(timeout: float | None = None) -> T ``` @@ -59,7 +56,6 @@ Raises: #### `result` ```python -@abstractmethod def result(timeout: float | None = None) -> T ``` diff --git a/docs/api/exceptions.md b/docs/api/exceptions.md index cbdecf2..97e9fdf 100644 --- a/docs/api/exceptions.md +++ b/docs/api/exceptions.md @@ -1,5 +1,3 @@ -# `tinker._exceptions` - ## `TinkerError` Objects ```python diff --git a/docs/api/restclient.md b/docs/api/restclient.md index 5bc7509..0a1b89f 100644 --- a/docs/api/restclient.md +++ b/docs/api/restclient.md @@ -1,5 +1,3 @@ -# `tinker.lib.public_interfaces.rest_client` - RestClient for Tinker API REST operations. ## `RestClient` Objects @@ -40,8 +38,6 @@ for checkpoint in checkpoints.checkpoints: #### `get_training_run` ```python -@sync_only -@capture_exceptions(fatal=True) def get_training_run( training_run_id: types.ModelID) -> ConcurrentFuture[types.TrainingRun] ``` @@ -64,7 +60,6 @@ print(f"Training Run ID: {response.training_run_id}, Base: {response.base_model} #### `get_training_run_async` ```python -@capture_exceptions(fatal=True) async def get_training_run_async( training_run_id: types.ModelID) -> types.TrainingRun ``` @@ -74,8 +69,6 @@ Async version of get_training_run. #### `get_training_run_by_tinker_path` ```python -@sync_only -@capture_exceptions(fatal=True) def get_training_run_by_tinker_path( tinker_path: str) -> ConcurrentFuture[types.TrainingRun] ``` @@ -98,7 +91,6 @@ print(f"Training Run ID: {response.training_run_id}, Base: {response.base_model} #### `get_training_run_by_tinker_path_async` ```python -@capture_exceptions(fatal=True) async def get_training_run_by_tinker_path_async( tinker_path: str) -> types.TrainingRun ``` @@ -108,7 +100,6 @@ Async version of get_training_run_by_tinker_path. #### `get_weights_info_by_tinker_path` ```python -@capture_exceptions(fatal=True) def get_weights_info_by_tinker_path( tinker_path: str) -> APIFuture[types.WeightsInfoResponse] ``` @@ -131,8 +122,6 @@ print(f"Base Model: {response.base_model}, LoRA Rank: {response.lora_rank}") #### `list_training_runs` ```python -@sync_only -@capture_exceptions(fatal=True) def list_training_runs( limit: int = 20, offset: int = 0) -> ConcurrentFuture[types.TrainingRunsResponse] @@ -160,7 +149,6 @@ next_page = rest_client.list_training_runs(limit=50, offset=50) #### `list_training_runs_async` ```python -@capture_exceptions(fatal=True) async def list_training_runs_async(limit: int = 20, offset: int = 0 ) -> types.TrainingRunsResponse @@ -171,8 +159,6 @@ Async version of list_training_runs. #### `list_checkpoints` ```python -@sync_only -@capture_exceptions(fatal=True) def list_checkpoints( training_run_id: types.ModelID ) -> ConcurrentFuture[types.CheckpointsListResponse] @@ -200,7 +186,6 @@ for checkpoint in response.checkpoints: #### `list_checkpoints_async` ```python -@capture_exceptions(fatal=True) async def list_checkpoints_async( training_run_id: types.ModelID) -> types.CheckpointsListResponse ``` @@ -210,8 +195,6 @@ Async version of list_checkpoints. #### `get_checkpoint_archive_url` ```python -@sync_only -@capture_exceptions(fatal=True) def get_checkpoint_archive_url( training_run_id: types.ModelID, checkpoint_id: str ) -> ConcurrentFuture[types.CheckpointArchiveUrlResponse] @@ -238,7 +221,6 @@ print(f"Expires at: {response.expires_at}") #### `get_checkpoint_archive_url_async` ```python -@capture_exceptions(fatal=True) async def get_checkpoint_archive_url_async( training_run_id: types.ModelID, checkpoint_id: str) -> types.CheckpointArchiveUrlResponse @@ -249,8 +231,6 @@ Async version of get_checkpoint_archive_url. #### `delete_checkpoint` ```python -@sync_only -@capture_exceptions(fatal=True) def delete_checkpoint(training_run_id: types.ModelID, checkpoint_id: str) -> ConcurrentFuture[None] ``` @@ -260,7 +240,6 @@ Delete a checkpoint for a training run. #### `delete_checkpoint_async` ```python -@capture_exceptions(fatal=True) async def delete_checkpoint_async(training_run_id: types.ModelID, checkpoint_id: str) -> None ``` @@ -270,8 +249,6 @@ Async version of delete_checkpoint. #### `delete_checkpoint_from_tinker_path` ```python -@sync_only -@capture_exceptions(fatal=True) def delete_checkpoint_from_tinker_path( tinker_path: str) -> ConcurrentFuture[None] ``` @@ -281,7 +258,6 @@ Delete a checkpoint referenced by a tinker path. #### `delete_checkpoint_from_tinker_path_async` ```python -@capture_exceptions(fatal=True) async def delete_checkpoint_from_tinker_path_async(tinker_path: str) -> None ``` @@ -290,8 +266,6 @@ Async version of delete_checkpoint_from_tinker_path. #### `get_checkpoint_archive_url_from_tinker_path` ```python -@sync_only -@capture_exceptions(fatal=True) def get_checkpoint_archive_url_from_tinker_path( tinker_path: str ) -> ConcurrentFuture[types.CheckpointArchiveUrlResponse] @@ -308,7 +282,6 @@ Returns: #### `get_checkpoint_archive_url_from_tinker_path_async` ```python -@capture_exceptions(fatal=True) async def get_checkpoint_archive_url_from_tinker_path_async( tinker_path: str) -> types.CheckpointArchiveUrlResponse ``` @@ -318,8 +291,6 @@ Async version of get_checkpoint_archive_url_from_tinker_path. #### `publish_checkpoint_from_tinker_path` ```python -@sync_only -@capture_exceptions(fatal=True) def publish_checkpoint_from_tinker_path( tinker_path: str) -> ConcurrentFuture[None] ``` @@ -351,7 +322,6 @@ print("Checkpoint published successfully") #### `publish_checkpoint_from_tinker_path_async` ```python -@capture_exceptions(fatal=True) async def publish_checkpoint_from_tinker_path_async(tinker_path: str) -> None ``` @@ -360,8 +330,6 @@ Async version of publish_checkpoint_from_tinker_path. #### `unpublish_checkpoint_from_tinker_path` ```python -@sync_only -@capture_exceptions(fatal=True) def unpublish_checkpoint_from_tinker_path( tinker_path: str) -> ConcurrentFuture[None] ``` @@ -393,7 +361,6 @@ print("Checkpoint unpublished successfully") #### `unpublish_checkpoint_from_tinker_path_async` ```python -@capture_exceptions(fatal=True) async def unpublish_checkpoint_from_tinker_path_async( tinker_path: str) -> None ``` @@ -403,8 +370,6 @@ Async version of unpublish_checkpoint_from_tinker_path. #### `list_user_checkpoints` ```python -@sync_only -@capture_exceptions(fatal=True) def list_user_checkpoints( limit: int = 100, offset: int = 0) -> ConcurrentFuture[types.CheckpointsListResponse] @@ -439,7 +404,6 @@ if response.cursor and response.cursor.offset + response.cursor.limit < response #### `list_user_checkpoints_async` ```python -@capture_exceptions(fatal=True) async def list_user_checkpoints_async(limit: int = 100, offset: int = 0 ) -> types.CheckpointsListResponse @@ -450,8 +414,6 @@ Async version of list_user_checkpoints. #### `get_session` ```python -@sync_only -@capture_exceptions(fatal=True) def get_session(session_id: str) -> ConcurrentFuture[types.GetSessionResponse] ``` @@ -474,7 +436,6 @@ print(f"Samplers: {len(response.sampler_ids)}") #### `get_session_async` ```python -@capture_exceptions(fatal=True) async def get_session_async(session_id: str) -> types.GetSessionResponse ``` @@ -483,8 +444,6 @@ Async version of get_session. #### `list_sessions` ```python -@sync_only -@capture_exceptions(fatal=True) def list_sessions( limit: int = 20, offset: int = 0) -> ConcurrentFuture[types.ListSessionsResponse] @@ -511,7 +470,6 @@ next_page = rest_client.list_sessions(limit=50, offset=50) #### `list_sessions_async` ```python -@capture_exceptions(fatal=True) async def list_sessions_async(limit: int = 20, offset: int = 0) -> types.ListSessionsResponse ``` @@ -521,7 +479,6 @@ Async version of list_sessions. #### `get_sampler` ```python -@capture_exceptions(fatal=True) def get_sampler(sampler_id: str) -> APIFuture[types.GetSamplerResponse] ``` @@ -549,7 +506,6 @@ print(f"Base model: {response.base_model}") #### `get_sampler_async` ```python -@capture_exceptions(fatal=True) async def get_sampler_async(sampler_id: str) -> types.GetSamplerResponse ``` diff --git a/docs/api/samplingclient.md b/docs/api/samplingclient.md index e7183ec..4970bd1 100644 --- a/docs/api/samplingclient.md +++ b/docs/api/samplingclient.md @@ -1,5 +1,3 @@ -# `tinker.lib.public_interfaces.sampling_client` - SamplingClient for Tinker API. ## `SamplingClient` Objects @@ -35,7 +33,6 @@ result = future.result() #### `sample` ```python -@capture_exceptions(fatal=True) def sample( prompt: types.ModelInput, num_samples: int, @@ -82,7 +79,6 @@ Async version of sample. #### `compute_logprobs` ```python -@capture_exceptions(fatal=True) def compute_logprobs( prompt: types.ModelInput) -> ConcurrentFuture[list[float | None]] ``` diff --git a/docs/api/serviceclient.md b/docs/api/serviceclient.md index da814d8..0895c2a 100644 --- a/docs/api/serviceclient.md +++ b/docs/api/serviceclient.md @@ -1,5 +1,3 @@ -# `tinker.lib.public_interfaces.service_client` - ServiceClient for Tinker API. ## `ServiceClient` Objects @@ -36,8 +34,6 @@ rest_client = client.create_rest_client() #### `get_server_capabilities` ```python -@sync_only -@capture_exceptions(fatal=True) def get_server_capabilities() -> types.GetServerCapabilitiesResponse ``` @@ -56,7 +52,6 @@ print(f"Max batch size: {capabilities.max_batch_size}") #### `get_server_capabilities_async` ```python -@capture_exceptions(fatal=True) async def get_server_capabilities_async( ) -> types.GetServerCapabilitiesResponse ``` @@ -66,8 +61,6 @@ Async version of get_server_capabilities. #### `create_lora_training_client` ```python -@sync_only -@capture_exceptions(fatal=True) def create_lora_training_client( base_model: str, rank: int = 32, @@ -106,7 +99,6 @@ training_client = service_client.create_lora_training_client( #### `create_lora_training_client_async` ```python -@capture_exceptions(fatal=True) async def create_lora_training_client_async( base_model: str, rank: int = 32, @@ -122,8 +114,6 @@ Async version of create_lora_training_client. #### `create_training_client_from_state` ```python -@sync_only -@capture_exceptions(fatal=True) def create_training_client_from_state( path: str, user_metadata: dict[str, str] | None = None) -> TrainingClient @@ -150,7 +140,6 @@ training_client = service_client.create_training_client_from_state( #### `create_training_client_from_state_async` ```python -@capture_exceptions(fatal=True) async def create_training_client_from_state_async( path: str, user_metadata: dict[str, str] | None = None) -> TrainingClient @@ -161,7 +150,6 @@ Async version of create_training_client_from_state. #### `create_sampling_client` ```python -@capture_exceptions(fatal=True) def create_sampling_client( model_path: str | None = None, base_model: str | None = None, @@ -197,7 +185,6 @@ sampling_client = service_client.create_sampling_client( #### `create_sampling_client_async` ```python -@capture_exceptions(fatal=True) async def create_sampling_client_async( model_path: str | None = None, base_model: str | None = None, @@ -209,7 +196,6 @@ Async version of create_sampling_client. #### `create_rest_client` ```python -@capture_exceptions(fatal=True) def create_rest_client() -> RestClient ``` diff --git a/docs/api/trainingclient.md b/docs/api/trainingclient.md index 1f5f420..b5e135b 100644 --- a/docs/api/trainingclient.md +++ b/docs/api/trainingclient.md @@ -1,5 +1,3 @@ -# `tinker.lib.public_interfaces.training_client` - TrainingClient for Tinker API. ## `TrainingClient` Objects @@ -34,7 +32,6 @@ sampling_client = training_client.save_weights_and_get_sampling_client("my-model #### `forward` ```python -@capture_exceptions(fatal=True) def forward( data: List[types.Datum], loss_fn: types.LossFnType, @@ -78,7 +75,6 @@ Async version of forward. #### `forward_backward` ```python -@capture_exceptions(fatal=True) def forward_backward( data: List[types.Datum], loss_fn: types.LossFnType, @@ -130,8 +126,6 @@ Async version of forward_backward. #### `forward_backward_custom` ```python -@sync_only -@capture_exceptions(fatal=True) def forward_backward_custom( data: List[types.Datum], loss_fn: CustomLossFnV1) -> APIFuture[types.ForwardBackwardOutput] @@ -166,7 +160,6 @@ print(f"Metrics: {result.metrics}") #### `forward_backward_custom_async` ```python -@capture_exceptions(fatal=True) async def forward_backward_custom_async( data: List[types.Datum], loss_fn: CustomLossFnV1) -> APIFuture[types.ForwardBackwardOutput] @@ -177,7 +170,6 @@ Async version of forward_backward_custom. #### `optim_step` ```python -@capture_exceptions(fatal=True) def optim_step( adam_params: types.AdamParams) -> APIFuture[types.OptimStepResponse] ``` @@ -220,7 +212,6 @@ Async version of optim_step. #### `save_state` ```python -@capture_exceptions(fatal=True) def save_state(name: str) -> APIFuture[types.SaveWeightsResponse] ``` @@ -251,7 +242,6 @@ Async version of save_state. #### `load_state` ```python -@capture_exceptions(fatal=True) def load_state(path: str) -> APIFuture[types.LoadWeightsResponse] ``` @@ -282,7 +272,6 @@ Async version of load_state. #### `load_state_with_optimizer` ```python -@capture_exceptions(fatal=True) def load_state_with_optimizer( path: str) -> APIFuture[types.LoadWeightsResponse] ``` @@ -317,7 +306,6 @@ Async version of load_state_with_optimizer. #### `save_weights_for_sampler` ```python -@capture_exceptions(fatal=True) def save_weights_for_sampler( name: str) -> APIFuture[types.SaveWeightsForSamplerResponse] ``` @@ -355,8 +343,6 @@ Async version of save_weights_for_sampler. #### `get_info` ```python -@sync_only -@capture_exceptions(fatal=True) def get_info() -> types.GetInfoResponse ``` @@ -376,7 +362,6 @@ print(f"LoRA rank: {info.model_data.lora_rank}") #### `get_info_async` ```python -@capture_exceptions(fatal=True) async def get_info_async() -> types.GetInfoResponse ``` @@ -385,8 +370,6 @@ Async version of get_info. #### `get_tokenizer` ```python -@cache -@capture_exceptions(fatal=True) def get_tokenizer() -> PreTrainedTokenizer ``` @@ -405,7 +388,6 @@ text = tokenizer.decode(tokens) #### `create_sampling_client` ```python -@capture_exceptions(fatal=True) def create_sampling_client( model_path: str, retry_config: RetryConfig | None = None) -> SamplingClient @@ -431,7 +413,6 @@ sampling_client = training_client.create_sampling_client( #### `create_sampling_client_async` ```python -@capture_exceptions(fatal=True) async def create_sampling_client_async( model_path: str, retry_config: RetryConfig | None = None) -> SamplingClient @@ -442,7 +423,6 @@ Async version of create_sampling_client. #### `save_weights_and_get_sampling_client` ```python -@capture_exceptions(fatal=True) def save_weights_and_get_sampling_client( name: str | None = None, retry_config: RetryConfig | None = None) -> SamplingClient @@ -471,7 +451,6 @@ result = sampling_client.sample(prompt, 1, params).result() #### `save_weights_and_get_sampling_client_async` ```python -@capture_exceptions(fatal=True) async def save_weights_and_get_sampling_client_async( name: str | None = None, retry_config: RetryConfig | None = None) -> SamplingClient diff --git a/docs/api/types.md b/docs/api/types.md index 5aaf180..05731a3 100644 --- a/docs/api/types.md +++ b/docs/api/types.md @@ -1,5 +1,3 @@ -# `tinker.types.optim_step_request` - ## `AdamParams` Objects ```python @@ -22,8 +20,6 @@ Coefficient used for computing running averages of gradient square Term added to the denominator to improve numerical stability -# `tinker.types.optim_step_response` - ## `OptimStepResponse` Objects ```python @@ -34,8 +30,6 @@ class OptimStepResponse(BaseModel) Optimization step metrics as key-value pairs -# `tinker.types.model_input` - ## `ModelInput` Objects ```python @@ -49,7 +43,6 @@ Sequence of input chunks (formerly TokenSequence) #### `from_ints` ```python -@classmethod def from_ints(cls, tokens: List[int]) -> "ModelInput" ``` @@ -67,7 +60,6 @@ Throws exception if there are any non-token chunks #### `length` ```python -@property def length() -> int ``` @@ -76,7 +68,6 @@ Return the total context length used by this ModelInput. #### `empty` ```python -@classmethod def empty(cls) -> "ModelInput" ``` @@ -98,8 +89,6 @@ def append_int(token: int) -> "ModelInput" Add a new token, return a new ModelInput. -# `tinker.types.weights_info_response` - ## `WeightsInfoResponse` Objects ```python @@ -108,8 +97,6 @@ class WeightsInfoResponse(BaseModel) Minimal information for loading public checkpoints. -# `tinker.types.checkpoint` - ## `Checkpoint` Objects ```python @@ -165,14 +152,11 @@ The checkpoint ID #### `from_tinker_path` ```python -@classmethod def from_tinker_path(cls, tinker_path: str) -> "ParsedCheckpointTinkerPath" ``` Parse a tinker path to an instance of ParsedCheckpointTinkerPath -# `tinker.types.checkpoint_archive_url_response` - ## `CheckpointArchiveUrlResponse` Objects ```python @@ -187,8 +171,6 @@ Signed URL to download the checkpoint archive Unix timestamp when the signed URL expires, if available -# `tinker.types.sampled_sequence` - ## `SampledSequence` Objects ```python @@ -207,8 +189,6 @@ List of generated token IDs Log probabilities for each token (optional) -# `tinker.types.try_again_response` - ## `TryAgainResponse` Objects ```python @@ -219,8 +199,6 @@ class TryAgainResponse(BaseModel) Request ID that is still pending -# `tinker.types.load_weights_request` - ## `LoadWeightsRequest` Objects ```python @@ -235,8 +213,6 @@ A tinker URI for model weights at a specific step Whether to load optimizer state along with model weights -# `tinker.types.telemetry_send_request` - ## `TelemetrySendRequest` Objects ```python @@ -251,8 +227,6 @@ Host platform name SDK version string -# `tinker.types.image_asset_pointer_chunk` - ## `ImageAssetPointerChunk` Objects ```python @@ -279,8 +253,6 @@ Number of tokens this image represents Image width in pixels -# `tinker.types.checkpoints_list_response` - ## `CheckpointsListResponse` Objects ```python @@ -295,8 +267,6 @@ List of available model checkpoints for the model Pagination cursor information (None for unpaginated responses) -# `tinker.types.generic_event` - ## `GenericEvent` Objects ```python @@ -319,8 +289,6 @@ Log severity level Arbitrary structured JSON payload -# `tinker.types.encoded_text_chunk` - ## `EncodedTextChunk` Objects ```python @@ -331,8 +299,6 @@ class EncodedTextChunk(StrictBase) Array of token IDs -# `tinker.types.forward_backward_input` - ## `ForwardBackwardInput` Objects ```python @@ -351,8 +317,6 @@ Fully qualified function path for the loss function Optional configuration parameters for the loss function (e.g., PPO clip thresholds, DPO beta) -# `tinker.types.session_start_event` - ## `SessionStartEvent` Objects ```python @@ -367,8 +331,6 @@ Telemetry event type Log severity level -# `tinker.types.training_runs_response` - ## `TrainingRunsResponse` Objects ```python @@ -383,8 +345,6 @@ List of training runs Pagination cursor information -# `tinker.types.save_weights_response` - ## `SaveWeightsResponse` Objects ```python @@ -395,8 +355,6 @@ class SaveWeightsResponse(BaseModel) A tinker URI for model weights at a specific step -# `tinker.types.sample_request` - ## `SampleRequest` Objects ```python @@ -444,8 +402,6 @@ Defaults to false. If set to a positive integer, returns the top-k logprobs for each prompt token. -# `tinker.types.forward_backward_output` - ## `ForwardBackwardOutput` Objects ```python @@ -464,8 +420,6 @@ Dictionary mapping field names to tensor data Training metrics as key-value pairs -# `tinker.types.sample_response` - ## `SampleResponse` Objects ```python @@ -485,8 +439,6 @@ the top-k logprobs are computed for every token in the prompt. The `topk_prompt_logprobs` response contains, for every token in the prompt, a list of up to k (token_id, logprob) tuples. -# `tinker.types.create_sampling_session_response` - ## `CreateSamplingSessionResponse` Objects ```python @@ -497,8 +449,6 @@ class CreateSamplingSessionResponse(BaseModel) The generated sampling session ID -# `tinker.types.cursor` - ## `Cursor` Objects ```python @@ -517,8 +467,6 @@ The maximum number of items requested The total number of items available -# `tinker.types.create_model_request` - ## `CreateModelRequest` Objects ```python @@ -529,8 +477,6 @@ class CreateModelRequest(StrictBase) Optional metadata about this model/training run, set by the end-user -# `tinker.types.datum` - ## `Datum` Objects ```python @@ -544,15 +490,11 @@ Dictionary mapping field names to tensor data #### `convert_tensors` ```python -@model_validator(mode="before") -@classmethod def convert_tensors(cls, data: Any) -> Any ``` Convert torch.Tensor and numpy arrays to TensorData in loss_fn_inputs during construction. -# `tinker.types.training_run` - ## `TrainingRun` Objects ```python @@ -599,8 +541,6 @@ The most recent sampler checkpoint, if available Optional metadata about this training run, set by the end-user -# `tinker.types.session_end_event` - ## `SessionEndEvent` Objects ```python @@ -619,8 +559,6 @@ Telemetry event type Log severity level -# `tinker.types.telemetry_batch` - ## `TelemetryBatch` Objects ```python @@ -635,8 +573,6 @@ Host platform name SDK version string -# `tinker.types.unhandled_exception_event` - ## `UnhandledExceptionEvent` Objects ```python @@ -655,8 +591,6 @@ Log severity level Optional Python traceback string -# `tinker.types.image_chunk` - ## `ImageChunk` Objects ```python @@ -693,8 +627,6 @@ match expected_tokens. #### `validate_data` ```python -@field_validator("data", mode="before") -@classmethod def validate_data(cls, value: Union[bytes, str]) -> bytes ``` @@ -703,14 +635,11 @@ Deserialize base64 string to bytes if needed. #### `serialize_data` ```python -@field_serializer("data") def serialize_data(value: bytes) -> str ``` Serialize bytes to base64 string for JSON. -# `tinker.types.save_weights_request` - ## `SaveWeightsRequest` Objects ```python @@ -721,8 +650,6 @@ class SaveWeightsRequest(StrictBase) A file/directory name for the weights -# `tinker.types.lora_config` - ## `LoraConfig` Objects ```python @@ -751,8 +678,6 @@ Whether to add loras to the MLP layers (including MoE layers) Whether to add loras to the attention layers -# `tinker.types.create_sampling_session_request` - ## `CreateSamplingSessionRequest` Objects ```python @@ -780,8 +705,6 @@ Optional tinker:// path to your model weights or LoRA weights. If not provided, samples against the base model. -# `tinker.types.future_retrieve_request` - ## `FutureRetrieveRequest` Objects ```python @@ -792,8 +715,6 @@ class FutureRetrieveRequest(StrictBase) The ID of the request to retrieve -# `tinker.types.tensor_data` - ## `TensorData` Objects ```python @@ -828,8 +749,6 @@ def to_torch() -> "torch.Tensor" Convert TensorData to torch tensor. -# `tinker.types.save_weights_for_sampler_request` - ## `SaveWeightsForSamplerRequest` Objects ```python @@ -840,8 +759,6 @@ class SaveWeightsForSamplerRequest(StrictBase) A file/directory name for the weights -# `tinker.types.sampling_params` - ## `SamplingParams` Objects ```python @@ -872,8 +789,6 @@ Top-k sampling parameter (-1 for no limit) Nucleus sampling probability -# `tinker.types.save_weights_for_sampler_response` - ## `SaveWeightsForSamplerResponseInternal` Objects ```python @@ -898,8 +813,6 @@ class SaveWeightsForSamplerResponse(BaseModel) A tinker URI for model weights for sampling at a specific step -# `tinker.types.load_weights_response` - ## `LoadWeightsResponse` Objects ```python diff --git a/pydoc-markdown.yml b/pydoc-markdown.yml index 663269f..2032ba8 100644 --- a/pydoc-markdown.yml +++ b/pydoc-markdown.yml @@ -21,5 +21,7 @@ renderer: code_lang: true escape_html_in_docstring: false insert_header_anchors: false - signature_code_block: true + render_module_header: false render_toc: false + signature_code_block: true + signature_with_decorators: false