# Aider chat conversation:
USER: add --openai-organization-id param that sets OpenAI(organization=...)
ASSISTANT: ```diff
--- aider/main.py
+++ aider/main.py
@@ ... @@
model_group.add_argument(
"--openai-api-deployment-id",
metavar="OPENAI_API_DEPLOYMENT_ID",
env_var="OPENAI_API_DEPLOYMENT_ID",
help="Specify the deployment_id",
)
+ model_group.add_argument(
+ "--openai-organization-id",
+ metavar="OPENAI_ORGANIZATION_ID",
+ env_var="OPENAI_ORGANIZATION_ID",
+ help="Specify the OpenAI organization ID",
+ )
model_group.add_argument(
"--openrouter",
dest="openai_api_base",
action="store_const",
@@ ... @@
if args.openai_api_type == "azure":
client = openai.AzureOpenAI(
api_key=args.openai_api_key,
azure_endpoint=args.openai_api_base,
api_version=args.openai_api_version,
azure_deployment=args.openai_api_deployment_id,
)
else:
kwargs = dict()
if args.openai_api_base:
kwargs["base_url"] = args.openai_api_base
if "openrouter.ai" in args.openai_api_base:
kwargs["default_headers"] = {
"HTTP-Referer": "http://aider.chat",
"X-Title": "Aider",
}
+ if args.openai_organization_id:
+ kwargs["organization"] = args.openai_organization_id
client = openai.OpenAI(api_key=args.openai_api_key, **kwargs)
@@ ... @@
if args.verbose:
show = scrub_sensitive_info(parser.format_values())
io.tool_output(show)
io.tool_output("Option settings:")
for arg, val in sorted(vars(args).items()):
io.tool_output(f" - {arg}: {scrub_sensitive_info(str(val))}")
+ if args.openai_organization_id:
+ io.tool_output(f" - openai_organization_id: {args.openai_organization_id}")
```
This diff adds a new command-line argument `--openai-organization-id` to the `aider/main.py` file, which allows the user to specify the OpenAI organization ID. It also ensures that if this argument is provided, it is passed to the `openai.OpenAI` client initialization as the `organization` parameter. Additionally, it outputs the organization ID if verbose mode is enabled.
# Aider chat conversation:
USER: add a meta canonical tag pointing to https://aider.chat/2023/11/06/benchmarks-speed-1106.html
ASSISTANT: ```diff
--- docs/benchmarks-speed-1106.md
+++ docs/benchmarks-speed-1106.md
@@ ... @@
title: Speed benchmarks of GPT-4 Turbo and gpt-3.5-turbo-1106
excerpt: This report provides a detailed comparison of the speed of GPT-4 Turbo and gpt-3.5-turbo-1106 models based on the aider benchmarking suite.
+<link rel="canonical" href="https://aider.chat/2023/11/06/benchmarks-speed-1106.html" />
highlight_image: /assets/benchmarks-speed-1106.svg
---
```
# Aider chat conversation:
USER: can the highlight image be bigger on mobile? does it make sense to use em not pixels to size it?
ASSISTANT: Using `em` units for the highlight image size can be beneficial for responsiveness, as `em` units are relative to the font-size of the element (or the body if not set), which can scale better across different screen sizes and maintain the design's typographic harmony.
To make the highlight image bigger on mobile, we can use a media query that targets smaller screen sizes and increases the `max-width` of the `.post-highlight` class in `em` units. Here's the diff for the changes:
```diff
--- assets/css/style.scss
+++ assets/css/style.scss
@@ ... @@
.post-highlight {
- max-width: 200px;
+ max-width: 12.5em; /* Assuming the base font-size is 16px, 12.5em is equivalent to 200px */
margin-right: 1em;
}
+@media (max-width: 768px) {
+ .post-highlight {
+ max-width: 18em; /* Larger size on mobile */
+ }
+}
+
.post-content {
display: flex;
align-items: flex-start;
```
This change assumes that the base font size is 16px, which is a common default for browsers. If the base font size is different, you may need to adjust the `em` values accordingly. The media query targets screens with a maximum width of 768px, which is a common breakpoint for tablets and mobile devices.