@@ -88,52 +88,7 @@ r = cli.responses.create(model="gpt-4.1", input="Hey!")
8888r
8989```
9090
91- Hey! How can I help you today? 😊
92-
93- <details >
94-
95- - id: resp_68b9978ecec48196aa3e77b09ed41c6403f00c61bc19c097
96- - created_at: 1756993423.0
97- - error: None
98- - incomplete_details: None
99- - instructions: None
100- - metadata: {}
101- - model: gpt-4.1-2025-04-14
102- - object: response
103- - output:
104- \[ ResponseOutputMessage(id=‘msg_68b9978f9f70819684b17b0f21072a9003f00c61bc19c097’,
105- content=\[ ResponseOutputText(annotations=\[\] , text=‘Hey! How can I
106- help you today? 😊’, type=‘output_text’, logprobs=\[\] )\] ,
107- role=‘assistant’, status=‘completed’, type=‘message’)\]
108- - parallel_tool_calls: True
109- - temperature: 1.0
110- - tool_choice: auto
111- - tools: \[\]
112- - top_p: 1.0
113- - background: False
114- - conversation: None
115- - max_output_tokens: None
116- - max_tool_calls: None
117- - previous_response_id: None
118- - prompt: None
119- - prompt_cache_key: None
120- - reasoning: Reasoning(effort=None, generate_summary=None, summary=None)
121- - safety_identifier: None
122- - service_tier: default
123- - status: completed
124- - text: ResponseTextConfig(format=ResponseFormatText(type=‘text’),
125- verbosity=‘medium’)
126- - top_logprobs: 0
127- - truncation: disabled
128- - usage: ResponseUsage(input_tokens=9,
129- input_tokens_details=InputTokensDetails(cached_tokens=0),
130- output_tokens=11,
131- output_tokens_details=OutputTokensDetails(reasoning_tokens=0),
132- total_tokens=20)
133- - user: None
134- - store: True
135-
136- </details >
91+ Response(id='resp_0f917a6452ee099400697b191473c48191b8e4f6e76e0add02', created_at=1769675028.0, error=None, incomplete_details=None, instructions=None, metadata={}, model='gpt-4.1-2025-04-14', object='response', output=[ResponseOutputMessage(id='msg_0f917a6452ee099400697b1914af1081919353425f8cd2d7d7', content=[ResponseOutputText(annotations=[], text='Hey! How can I help you today? 😊', type='output_text', logprobs=[])], role='assistant', status='completed', type='message')], parallel_tool_calls=True, temperature=1.0, tool_choice='auto', tools=[], top_p=1.0, background=False, completed_at=1769675028.0, conversation=None, max_output_tokens=None, max_tool_calls=None, previous_response_id=None, prompt=None, prompt_cache_key=None, prompt_cache_retention=None, reasoning=Reasoning(effort=None, generate_summary=None, summary=None), safety_identifier=None, service_tier='default', status='completed', text=ResponseTextConfig(format=ResponseFormatText(type='text'), verbosity='medium'), top_logprobs=0, truncation='disabled', usage=ResponseUsage(input_tokens=9, input_tokens_details=InputTokensDetails(cached_tokens=0), output_tokens=11, output_tokens_details=OutputTokensDetails(reasoning_tokens=0), total_tokens=20), user=None, billing={'payer': 'developer'}, frequency_penalty=0.0, presence_penalty=0.0, store=True)
13792
13893If you run the same request again it will read it from the cache.
13994
@@ -142,49 +97,4 @@ r = cli.responses.create(model="gpt-4.1", input="Hey!")
14297r
14398```
14499
145- Hey! How can I help you today? 😊
146-
147- <details >
148-
149- - id: resp_68b9978ecec48196aa3e77b09ed41c6403f00c61bc19c097
150- - created_at: 1756993423.0
151- - error: None
152- - incomplete_details: None
153- - instructions: None
154- - metadata: {}
155- - model: gpt-4.1-2025-04-14
156- - object: response
157- - output:
158- \[ ResponseOutputMessage(id=‘msg_68b9978f9f70819684b17b0f21072a9003f00c61bc19c097’,
159- content=\[ ResponseOutputText(annotations=\[\] , text=‘Hey! How can I
160- help you today? 😊’, type=‘output_text’, logprobs=\[\] )\] ,
161- role=‘assistant’, status=‘completed’, type=‘message’)\]
162- - parallel_tool_calls: True
163- - temperature: 1.0
164- - tool_choice: auto
165- - tools: \[\]
166- - top_p: 1.0
167- - background: False
168- - conversation: None
169- - max_output_tokens: None
170- - max_tool_calls: None
171- - previous_response_id: None
172- - prompt: None
173- - prompt_cache_key: None
174- - reasoning: Reasoning(effort=None, generate_summary=None, summary=None)
175- - safety_identifier: None
176- - service_tier: default
177- - status: completed
178- - text: ResponseTextConfig(format=ResponseFormatText(type=‘text’),
179- verbosity=‘medium’)
180- - top_logprobs: 0
181- - truncation: disabled
182- - usage: ResponseUsage(input_tokens=9,
183- input_tokens_details=InputTokensDetails(cached_tokens=0),
184- output_tokens=11,
185- output_tokens_details=OutputTokensDetails(reasoning_tokens=0),
186- total_tokens=20)
187- - user: None
188- - store: True
189-
190- </details >
100+ Response(id='resp_0f917a6452ee099400697b191473c48191b8e4f6e76e0add02', created_at=1769675028.0, error=None, incomplete_details=None, instructions=None, metadata={}, model='gpt-4.1-2025-04-14', object='response', output=[ResponseOutputMessage(id='msg_0f917a6452ee099400697b1914af1081919353425f8cd2d7d7', content=[ResponseOutputText(annotations=[], text='Hey! How can I help you today? 😊', type='output_text', logprobs=[])], role='assistant', status='completed', type='message')], parallel_tool_calls=True, temperature=1.0, tool_choice='auto', tools=[], top_p=1.0, background=False, completed_at=1769675028.0, conversation=None, max_output_tokens=None, max_tool_calls=None, previous_response_id=None, prompt=None, prompt_cache_key=None, prompt_cache_retention=None, reasoning=Reasoning(effort=None, generate_summary=None, summary=None), safety_identifier=None, service_tier='default', status='completed', text=ResponseTextConfig(format=ResponseFormatText(type='text'), verbosity='medium'), top_logprobs=0, truncation='disabled', usage=ResponseUsage(input_tokens=9, input_tokens_details=InputTokensDetails(cached_tokens=0), output_tokens=11, output_tokens_details=OutputTokensDetails(reasoning_tokens=0), total_tokens=20), user=None, billing={'payer': 'developer'}, frequency_penalty=0.0, presence_penalty=0.0, store=True)
0 commit comments