Skip to content

Commit 38de55f

Browse files
committed
fix links
2 parents 0e66745 + 484399b commit 38de55f

File tree

13 files changed

+180
-214
lines changed

13 files changed

+180
-214
lines changed

docs/cloud/bestpractices/avoidoptimizefinal.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -20,10 +20,10 @@ It is important to note that using this optimization will force a rewrite of a p
2020
even if merging to a single part has already occurred.
2121

2222
Additionally, use of the `OPTIMIZE TABLE ... FINAL` query may disregard
23-
setting [`max_bytes_to_merge_at_max_space_in_pool`](/operations/settings/merge-tree-settings#max-bytes-to-merge-at-max-space-in-pool) which controls the maximum size of parts
23+
setting [`max_bytes_to_merge_at_max_space_in_pool`](/operations/settings/merge-tree-settings#max_bytes_to_merge_at_max_space_in_pool) which controls the maximum size of parts
2424
that ClickHouse will typically merge by itself in the background.
2525

26-
The [`max_bytes_to_merge_at_max_space_in_pool`](/operations/settings/merge-tree-settings#max-bytes-to-merge-at-max-space-in-pool) setting is by default set to 150 GB.
26+
The [`max_bytes_to_merge_at_max_space_in_pool`](/operations/settings/merge-tree-settings#max_bytes_to_merge_at_max_space_in_pool) setting is by default set to 150 GB.
2727
When running `OPTIMIZE TABLE ... FINAL`,
2828
the steps outlined above will be performed resulting in a single part after merge.
2929
This remaining single part could exceed the 150 GB specified by the default of this setting.

docs/guides/developer/deduplicating-inserts-on-retries.md

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -15,7 +15,7 @@ When an insert is retried, ClickHouse tries to determine whether the data has al
1515

1616
**Only `*MergeTree` engines support deduplication on insertion.**
1717

18-
For `*ReplicatedMergeTree` engines, insert deduplication is enabled by default and is controlled by the [`replicated_deduplication_window`](/operations/settings/merge-tree-settings#replicated-deduplication-window) and [`replicated_deduplication_window_seconds`](/operations/settings/merge-tree-settings#replicated-deduplication-window-seconds) settings. For non-replicated `*MergeTree` engines, deduplication is controlled by the [`non_replicated_deduplication_window`](/operations/settings/merge-tree-settings#non-replicated-deduplication-window) setting.
18+
For `*ReplicatedMergeTree` engines, insert deduplication is enabled by default and is controlled by the [`replicated_deduplication_window`](/operations/settings/merge-tree-settings#replicated_deduplication_window) and [`replicated_deduplication_window_seconds`](/operations/settings/merge-tree-settings#replicated_deduplication_window_seconds) settings. For non-replicated `*MergeTree` engines, deduplication is controlled by the [`non_replicated_deduplication_window`](/operations/settings/merge-tree-settings#non_replicated_deduplication_window) setting.
1919

2020
The settings above determine the parameters of the deduplication log for a table. The deduplication log stores a finite number of `block_id`s, which determine how deduplication works (see below).
2121

@@ -41,9 +41,9 @@ When a table has one or more materialized views, the inserted data is also inser
4141

4242
You can control this process using the following settings for the source table:
4343

44-
- [`replicated_deduplication_window`](/operations/settings/merge-tree-settings#replicated-deduplication-window)
45-
- [`replicated_deduplication_window_seconds`](/operations/settings/merge-tree-settings#replicated-deduplication-window-seconds)
46-
- [`non_replicated_deduplication_window`](/operations/settings/merge-tree-settings#non-replicated-deduplication-window)
44+
- [`replicated_deduplication_window`](/operations/settings/merge-tree-settings#replicated_deduplication_window)
45+
- [`replicated_deduplication_window_seconds`](/operations/settings/merge-tree-settings#replicated_deduplication_window_seconds)
46+
- [`non_replicated_deduplication_window`](/operations/settings/merge-tree-settings#non_replicated_deduplication_window)
4747

4848
You can also use the user profile setting [`deduplicate_blocks_in_dependent_materialized_views`](/operations/settings/settings#deduplicate_blocks_in_dependent_materialized_views).
4949

docs/guides/developer/time-series-filling-gaps.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -301,7 +301,7 @@ INTERPOLATE (cumulative);
301301
```
302302

303303
That looks much better.
304-
And now to finish it off, let's add a bar chart using the `bar` function, not forgetting to add our new column to the `INTERPPOLATE` clause.
304+
And now to finish it off, let's add a bar chart using the `bar` function, not forgetting to add our new column to the `INTERPOLATE` clause.
305305

306306
```sql
307307
SELECT

docs/integrations/data-ingestion/dbms/postgresql/connecting-to-postgresql.md

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -12,15 +12,15 @@ import ExperimentalBadge from '@theme/badges/ExperimentalBadge';
1212

1313
This page covers following options for integrating PostgreSQL with ClickHouse:
1414

15-
- using [ClickPipes](/integrations/clickpipes/postgres), the managed integration service for ClickHouse Cloud - now in Private Preview. Please [sign up here](https://clickpipes.peerdb.io/)
15+
- using [ClickPipes](/integrations/clickpipes/postgres), the managed integration service for ClickHouse Cloud - now in public beta. Please [sign up here](https://clickpipes.peerdb.io/)
1616
- using `PeerDB by ClickHouse`, a CDC tool specifically designed for PostgreSQL database replication to both self-hosted ClickHouse and ClickHouse Cloud
17-
- PeerDB is now available natively in ClickHouse Cloud - Blazing-fast Postgres to ClickHouse CDC with our [new ClickPipe connector](/integrations/clickpipes/postgres) - now in Private Preview. Please [sign up here](https://clickpipes.peerdb.io/)
17+
- PeerDB is now available natively in ClickHouse Cloud - Blazing-fast Postgres to ClickHouse CDC with our [new ClickPipe connector](/integrations/clickpipes/postgres) - now in public beta. Please [sign up here](https://clickpipes.peerdb.io/)
1818
- using the `PostgreSQL` table engine, for reading from a PostgreSQL table
1919
- using the experimental `MaterializedPostgreSQL` database engine, for syncing a database in PostgreSQL with a database in ClickHouse
2020

2121
## Using ClickPipes (powered by PeerDB) {#using-clickpipes-powered-by-peerdb}
2222

23-
PeerDB is now available natively in ClickHouse Cloud - Blazing-fast Postgres to ClickHouse CDC with our [new ClickPipe connector](/integrations/clickpipes/postgres) - now in Private Preview. Please [sign up here](https://clickpipes.peerdb.io/)
23+
PeerDB is now available natively in ClickHouse Cloud - Blazing-fast Postgres to ClickHouse CDC with our [new ClickPipe connector](/integrations/clickpipes/postgres) - now in public beta. Please [sign up here](https://clickpipes.peerdb.io/)
2424

2525
## Using the PostgreSQL Table Engine {#using-the-postgresql-table-engine}
2626

docs/integrations/data-ingestion/dbms/postgresql/inserting-data.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -9,8 +9,8 @@ We recommend reading [this guide](/guides/inserting-data) to learn best practice
99

1010
For bulk loading data from PostgreSQL, users can use:
1111

12-
- using [ClickPipes](/integrations/clickpipes/postgres), the managed integration service for ClickHouse Cloud - now in Private Preview. Please [sign up here](https://clickpipes.peerdb.io/)
12+
- using [ClickPipes](/integrations/clickpipes/postgres), the managed integration service for ClickHouse Cloud - now in public beta. Please [sign up here](https://clickpipes.peerdb.io/)
1313
- `PeerDB by ClickHouse`, an ETL tool specifically designed for PostgreSQL database replication to both self-hosted ClickHouse and ClickHouse Cloud.
14-
- PeerDB is now available natively in ClickHouse Cloud - Blazing-fast Postgres to ClickHouse CDC with our [new ClickPipe connector](/integrations/clickpipes/postgres) - now in Private Preview. Please [sign up here](https://clickpipes.peerdb.io/)
14+
- PeerDB is now available natively in ClickHouse Cloud - Blazing-fast Postgres to ClickHouse CDC with our [new ClickPipe connector](/integrations/clickpipes/postgres) - now in public beta. Please [sign up here](https://clickhouse.com/cloud/clickpipes/postgres-cdc-connector)
1515
- The [Postgres Table Function](/sql-reference/table-functions/postgresql) to read data directly. This is typically appropriate for if batch replication based on a known watermark, e.g. a timestamp. is sufficient or if it's a once-off migration. This approach can scale to 10's of millions of rows. Users looking to migrate larger datasets should consider multiple requests, each dealing with a chunk of the data. Staging tables can be used for each chunk prior to its partitions being moved to a final table. This allows failed requests to be retried. For further details on this bulk-loading strategy, see here.
1616
- Data can be exported from Postgres in CSV format. This can then be inserted into ClickHouse from either local files or via object storage using table functions.

docs/integrations/data-visualization/embeddable-and-clickhouse.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -13,7 +13,7 @@ import CommunityMaintainedBadge from '@theme/badges/CommunityMaintained';
1313

1414
<CommunityMaintainedBadge/>
1515

16-
In [Embeddable](https://embeddable.com/) you define [Data Models](https://trevorio.notion.site/Data-modeling-35637bbbc01046a1bc47715456bfa1d8) and [Components](https://trevorio.notion.site/Using-components-761f52ac2d0743b488371088a1024e49) in code (stored in your own code repository) and use our **SDK** to make these available for your team in the powerful Embeddable **no-code builder.**
16+
In [Embeddable](https://embeddable.com/) you define [Data Models](https://docs.embeddable.com/data-modeling/introduction) and [Components](https://docs.embeddable.com/development/introduction) in code (stored in your own code repository) and use our **SDK** to make these available for your team in the powerful Embeddable **no-code builder.**
1717

1818
The end result is the ability to deliver fast, interactive customer-facing analytics directly in your product; designed by your product team; built by your engineering team; maintained by your customer-facing and data teams. Exactly the way it should be.
1919

@@ -68,4 +68,4 @@ The `credentials` is a JavaScript object containing the necessary credentials ex
6868
- These are securely encrypted and only used to retrieve exactly the data you have described in your data models.
6969
Embeddable strongly encourage you to create a read-only database user for each connection (Embeddable will only ever read from your database, not write).
7070

71-
In order to support connecting to different databases for prod, qa, test, etc (or to support different databases for different customers) you can assign each connection to an environment (see [Environments API](https://www.notion.so/Environments-API-497169036b5148b38f7936aa75e62949?pvs=21)).
71+
In order to support connecting to different databases for prod, qa, test, etc (or to support different databases for different customers) you can assign each connection to an environment (see [Environments API](https://docs.embeddable.com/data/environments)).

docusaurus.config.en.js

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -56,10 +56,10 @@ const config = {
5656
// url: process.env.VERCEL_URL ? `https://${process.env.VERCEL_URL}` : 'https://bookish-disco-5997zvo.pages.github.io',
5757
baseUrl: "/docs/",
5858
baseUrlIssueBanner: true,
59-
onBrokenLinks: "warn",
59+
onBrokenLinks: "throw",
6060
onBrokenMarkdownLinks: "warn",
6161
onDuplicateRoutes: "throw",
62-
onBrokenAnchors: "warn",
62+
onBrokenAnchors: "throw",
6363
favicon: "img/docs_favicon.ico",
6464
organizationName: "ClickHouse",
6565
trailingSlash: false,

scripts/search/run_indexer.sh

Lines changed: 0 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -38,14 +38,12 @@ else
3838
fi
3939

4040
BASE_DIRECTORY_ARG="$BASE_DIRECTORY"
41-
SUB_DIRECTORIES="docs/en,knowledgebase"
4241
DRY_RUN=false
4342

4443
# allows us to override params if needed
4544
while [[ "$#" -gt 0 ]]; do
4645
case $1 in
4746
-d|--base_directory) BASE_DIRECTORY_ARG="$2"; shift ;;
48-
-s|--sub_directories) SUB_DIRECTORIES="$2"; shift ;;
4947
-x|--dry_run) DRY_RUN=true ;;
5048
*) echo "Unknown parameter passed: $1"; exit 1 ;;
5149
esac
@@ -56,7 +54,6 @@ done
5654
echo "Running the Python script..."
5755
python "$PYTHON_SCRIPT" \
5856
--base_directory "$BASE_DIRECTORY_ARG" \
59-
--sub_directories "$SUB_DIRECTORIES" \
6057
$( [ "$DRY_RUN" = true ] && echo "--dry_run" ) \
6158
--algolia_app_id "$ALGOLIA_APP_ID" \
6259
--algolia_api_key "$ALGOLIA_API_KEY"

scripts/settings/autogenerate-settings.sh

Lines changed: 8 additions & 195 deletions
Original file line numberDiff line numberDiff line change
@@ -16,9 +16,9 @@ else
1616
exit 1
1717
fi
1818

19-
2019
target_dir=$(dirname "$(dirname "$(realpath "$0")")")
2120
SCRIPT_NAME=$(basename "$0")
21+
SCRIPT_DIR="$( cd "$( dirname "${BASH_SOURCE[0]}" )" && pwd )"
2222
tmp_dir="$target_dir/tmp"
2323

2424
mkdir -p "$tmp_dir" || exit 1
@@ -41,204 +41,17 @@ fi
4141

4242
echo "Downloaded to: $script_path"
4343
echo "[$SCRIPT_NAME] Auto-generating settings"
44-
45-
# Autogenerate Format settings
4644
chmod +x "$script_path" || { echo "Error: Failed to set execute permission"; exit 1; }
47-
4845
root=$(dirname "$(dirname "$(realpath "$tmp_dir")")")
4946

50-
./clickhouse -q "
51-
WITH
52-
'FormatFactorySettings.h' AS cpp_file,
53-
settings_from_cpp AS
54-
(
55-
SELECT extract(line, 'DECLARE\\(\\w+, (\\w+),') AS name
56-
FROM file(cpp_file, LineAsString)
57-
WHERE match(line, '^\\s*DECLARE\\(')
58-
),
59-
main_content AS
60-
(
61-
SELECT format('## {} {}\\n{}\\n\\nType: \`{}\`\\n\\nDefault: \`{}\`\\n\\n{}\\n\\n',
62-
name, '{#'||name||'}', multiIf(tier == 'Experimental', '<ExperimentalBadge/>', tier == 'Beta', '<BetaBadge/>', ''), type, default, trim(BOTH '\\n' FROM description))
63-
FROM system.settings WHERE name IN settings_from_cpp
64-
ORDER BY name
65-
),
66-
'---
67-
title: ''Format Settings''
68-
sidebar_label: ''Format Settings''
69-
slug: /operations/settings/formats
70-
toc_max_heading_level: 2
71-
description: ''Settings which control input and output formats.''
72-
---
73-
74-
import ExperimentalBadge from \'@theme/badges/ExperimentalBadge\';
75-
import BetaBadge from \'@theme/badges/BetaBadge\';
76-
77-
<!-- Autogenerated -->
78-
These settings are autogenerated from [source](https://github.com/ClickHouse/ClickHouse/blob/master/src/Core/FormatFactorySettings.h).
79-
80-
' AS prefix
81-
SELECT prefix || (SELECT groupConcat(*) FROM main_content)
82-
INTO OUTFILE 'settings-formats.md' TRUNCATE FORMAT LineAsString
83-
" > /dev/null || { echo "Failed to Autogenerate Format settings"; exit 1; }
84-
85-
# Autogenerate settings
86-
./clickhouse -q "
87-
WITH
88-
'Settings.cpp' AS cpp_file,
89-
settings_from_cpp AS
90-
(
91-
SELECT extract(line, 'DECLARE\\(\\w+, (\\w+),') AS name
92-
FROM file(cpp_file, LineAsString)
93-
WHERE match(line, '^\\s*DECLARE\\(')
94-
),
95-
main_content AS
96-
(
97-
SELECT format(
98-
'## {} {}\\n{}\\n{}\\n\\nType: {}\\n\\nDefault value: {}\\n\\n{}\\n\\n',
99-
name,
100-
'{#'||name||'}',
101-
multiIf(tier == 'Experimental', '<ExperimentalBadge/>', tier == 'Beta', '<BetaBadge/>', ''),
102-
if(description LIKE '%Only has an effect in ClickHouse Cloud%', '\\n<CloudAvailableBadge/>', ''),
103-
type,
104-
default,
105-
replaceOne(
106-
trim(BOTH '\\n' FROM description),
107-
' and [MaterializedMySQL](../../engines/database-engines/materialized-mysql.md)',''
108-
)
109-
)
110-
FROM system.settings WHERE name IN settings_from_cpp
111-
ORDER BY name
112-
),
113-
'---
114-
title: ''Session Settings''
115-
sidebar_label: ''Session Settings''
116-
slug: /operations/settings/settings
117-
toc_max_heading_level: 2
118-
description: ''Settings which are found in the ``system.settings`` table.''
119-
---
120-
121-
import ExperimentalBadge from \'@theme/badges/ExperimentalBadge\';
122-
import BetaBadge from \'@theme/badges/BetaBadge\';
123-
import CloudAvailableBadge from \'@theme/badges/CloudAvailableBadge\';
124-
125-
<!-- Autogenerated -->
126-
All below settings are also available in table [system.settings](/docs/operations/system-tables/settings). These settings are autogenerated from [source](https://github.com/ClickHouse/ClickHouse/blob/master/src/Core/Settings.cpp).
127-
128-
' AS prefix
129-
SELECT prefix || (SELECT groupConcat(*) FROM main_content)
130-
INTO OUTFILE 'settings.md' TRUNCATE FORMAT LineAsString
131-
" > /dev/null || { echo "Failed to Autogenerate Core settings"; exit 1; }
132-
133-
# Autogenerate MergeTree settings
134-
./clickhouse -q "
135-
WITH
136-
merge_tree_settings AS
137-
(
138-
SELECT format(
139-
'## {} {} \n{}\nType: \`{}\`\n\nDefault: \`{}\`\n{}',
140-
name,
141-
'{#'||name||'}',
142-
multiIf(tier == 'Experimental', '\n<ExperimentalBadge/>\n', tier == 'Beta', '\n<BetaBadge/>\n', ''),
143-
type,
144-
default,
145-
replaceRegexpAll(description, '(?m)(^[ \t]+|[ \t]+$)', '')
146-
)
147-
FROM system.merge_tree_settings ORDER BY name
148-
)
149-
SELECT * FROM merge_tree_settings
150-
INTO OUTFILE 'generated_merge_tree_settings.md' TRUNCATE FORMAT LineAsString
151-
" > /dev/null || { echo "Failed to Autogenerate Core settings"; exit 1; }
152-
153-
# Auto generate global server settings
154-
./clickhouse -q "
155-
WITH
156-
server_settings_outside_source AS
157-
(
158-
SELECT
159-
arrayJoin(extractAllGroups(raw_blob, '## (\\w+)(?:\\s[^\n]+)?\n\\s+((?:[^#]|#[^#]|##[^ ])+)')) AS g,
160-
g[1] AS name,
161-
replaceRegexpAll(replaceRegexpAll(g[2], '\n(Type|Default( value)?): [^\n]+\n', ''), '^\n+|\n+$', '') AS doc
162-
FROM file('_server_settings_outside_source.md', RawBLOB)
163-
),
164-
server_settings_in_source AS
165-
(
166-
SELECT
167-
name,
168-
replaceRegexpAll(description, '(?m)^[ \t]+', '') AS description,
169-
type,
170-
default
171-
FROM system.server_settings
172-
),
173-
combined_server_settings AS
174-
(
175-
SELECT
176-
name,
177-
description,
178-
type,
179-
default
180-
FROM server_settings_in_source
181-
UNION ALL
182-
SELECT
183-
name,
184-
doc AS description,
185-
'' AS type,
186-
'' AS default
187-
FROM server_settings_outside_source
188-
),
189-
formatted_settings AS
190-
(
191-
SELECT
192-
format(
193-
'## {} {}\n\n{}{}{}\n\n',
194-
name,
195-
lcase('{#'||name||'}'),
196-
if(type != '', concat('Type: \`', type, '\`\n\n'), ''),
197-
if(default != '', concat('Default: \`', default, '\`\n\n'), ''),
198-
description
199-
) AS formatted_text
200-
FROM combined_server_settings
201-
ORDER BY name ASC
202-
),
203-
prefix_text AS
204-
(
205-
SELECT
206-
'---
207-
description: ''This section contains descriptions of server settings i.e settings
208-
which cannot be changed at the session or query level.''
209-
keywords: [''global server settings'']
210-
sidebar_label: ''Server Settings''
211-
sidebar_position: 57
212-
slug: /operations/server-configuration-parameters/settings
213-
title: ''Server Settings''
214-
---
215-
216-
import Tabs from ''@theme/Tabs'';
217-
import TabItem from ''@theme/TabItem'';
218-
import SystemLogParameters from ''@site/docs/operations/server-configuration-parameters/_snippets/_system-log-parameters.md''
219-
220-
# Server Settings
221-
222-
This section contains descriptions of server settings. These are settings which
223-
cannot be changed at the session or query level.
224-
225-
For more information on configuration files in ClickHouse see [""Configuration Files""](/operations/configuration-files).
226-
227-
Other settings are described in the ""[Settings](/operations/settings/overview)"" section.
228-
Before studying the settings, we recommend reading the [Configuration files](/operations/configuration-files)
229-
section and note the use of substitutions (the `incl` and `optional` attributes).
230-
231-
' AS prefix_content
232-
)
233-
SELECT
234-
arrayStringConcat([
235-
(SELECT prefix_content FROM prefix_text),
236-
arrayStringConcat(groupArray(formatted_text), '')
237-
], '')
238-
FROM formatted_settings
239-
INTO OUTFILE 'server_settings.md'
240-
TRUNCATE FORMAT LineAsString" > /dev/null || { echo "Failed to Autogenerate Format settings"; exit 1; }
47+
# Autogenerate settings for all .sql files in directory
48+
for SQL_FILE in "$SCRIPT_DIR"/*.sql; do
49+
if [ -f "$SQL_FILE" ]; then
50+
./clickhouse --queries-file "$SQL_FILE" > /dev/null || { echo "Failed to generate some settings"; exit 1; }
51+
fi
52+
done
24153

54+
# move across files to where they need to be
24255
mv settings-formats.md "$root/docs/operations/settings" || { echo "Failed to move generated settings-format.md"; exit 1; }
24356
mv settings.md "$root/docs/operations/settings" || { echo "Failed to move generated settings.md"; exit 1; }
24457
cat generated_merge_tree_settings.md >> "$root/docs/operations/settings/merge-tree-settings.md" || { echo "Failed to append MergeTree settings.md"; exit 1; }

0 commit comments

Comments
 (0)