Primary Key Continue When Row Deleted
On this page
The DELETE
statement deletes rows from a table.
Required privileges
The user must have the DELETE
and SELECT
privileges on the table.
Synopsis
Parameters
Parameter | Description |
---|---|
common_table_expr | See Common Table Expressions. |
table_name | The name of the table that contains the rows you want to update. |
AS table_alias_name | An alias for the table name. When an alias is provided, it completely hides the actual table name. |
WHERE a_expr | a_expr must be an expression that returns Boolean values using columns (e.g., <column> = <value> ). Delete rows that return TRUE . Without a |
sort_clause | An ORDER BY clause. See Ordering of rows in DML statements for more details. |
limit_clause | A LIMIT clause. See Limit Query Results for more details. |
RETURNING target_list | Return values based on rows deleted, where target_list can be specific column names from the table, * for all columns, or computations using scalar expressions. To return nothing in the response, not even the number of rows updated, use |
ONLY ... * | Supported for compatibility with PostgreSQL table inheritance syntax. This clause is a no-op, as CockroachDB does not currently support table inheritance. |
Success responses
Successful DELETE
statements return one of the following:
Response | Description |
---|---|
DELETE int | int rows were deleted. |
Retrieved table | Including the RETURNING clause retrieves the deleted rows, using the columns identified by the clause's parameters.See an example. |
Disk space usage after deletes
Deleting a row does not immediately free up the disk space. This is due to the fact that CockroachDB retains the ability to query tables historically.
If disk usage is a concern, the solution is to reduce the time-to-live (TTL) for the zone by setting gc.ttlseconds
to a lower value, which will cause garbage collection to clean up deleted objects (rows, tables) more frequently.
Select performance on deleted rows
Queries that scan across tables that have lots of deleted rows will have to scan over deletions that have not yet been garbage collected. Certain database usage patterns that frequently scan over and delete lots of rows will want to reduce the time-to-live values to clean up deleted rows more frequently.
Sorting the output of deletes
To sort the output of a DELETE
statement, use:
> WITH a AS ( DELETE ... RETURNING ...) SELECT ... FROM a ORDER BY ...
For an example, see Sort and return deleted rows.
For more information about ordering query results in general, see Ordering Query Results and Ordering of rows in DML statements.
Force index selection for deletes
By using the explicit index annotation (also known as "index hinting"), you can override CockroachDB's index selection and use a specific index for deleting rows of a named table.
Note:
Index selection can impact performance, but does not change the result of a query.
The syntax to force a specific index for a delete is:
> DELETE FROM table @ my_idx ;
This is equivalent to the longer expression:
> DELETE FROM table @ { FORCE_INDEX = my_idx };
To view how the index hint modifies the query plan that CockroachDB follows for deleting rows, use an EXPLAIN
statement. To see all indexes available on a table, use SHOW INDEXES
.
For examples, see Delete with index hints.
You can use the @primary
alias to use the table's primary key in your query if no secondary index explicitly named primary
exists on that table.
Preserving DELETE
performance over time
CockroachDB relies on multi-version concurrency control (MVCC) to process concurrent requests while guaranteeing strong consistency. As such, when you delete a row, it is not immediately removed from disk. The MVCC values for the row will remain until the garbage collection period defined by the gc.ttlseconds
variable in the applicable zone configuration has passed. By default, this period is 25 hours.
This means that with the default settings, each iteration of your DELETE
statement must scan over all of the rows previously marked for deletion within the last 25 hours. If you try to delete 10,000 rows 10 times within the same 25 hour period, the 10th command will have to scan over the 90,000 rows previously marked for deletion.
To preserve performance over iterative DELETE
queries, we recommend taking one of the following approaches:
- At each iteration, update the
WHERE
clause to filter only the rows that have not yet been marked for deletion. For an example, see Batch-delete on an indexed filter. - At each iteration, first use a
SELECT
statement to return primary key values on rows that are not yet deleted. Rows marked for deletion will not be returned. Then, use a nestedDELETE
loop over a smaller batch size, filtering on the primary key values. For an example, see Batch delete on a non-indexed column. - To iteratively delete rows in constant time, using a simple
DELETE
loop, you can alter your zone configuration and changegc.ttlseconds
to a low value like 5 minutes (i.e.,300
), and then run yourDELETE
statement once per GC interval.
Examples
Setup
The following examples use MovR, a fictional vehicle-sharing application, to demonstrate CockroachDB SQL statements. For more information about the MovR example application and dataset, see MovR: A Global Vehicle-sharing App.
To follow along, run cockroach demo
to start a temporary, in-memory cluster with the movr
dataset preloaded:
Delete rows using Primary Key/unique columns
Using columns with the Primary Key or Unique constraints to delete rows ensures your statement is unambiguous—no two rows contain the same column value, so it's less likely to delete data unintentionally.
In this example, code
is our primary key and we want to delete the row where the code equals "about_stuff_city". Because we're positive no other rows have that value in the code
column, there's no risk of accidentally removing another row.
> DELETE FROM promo_codes WHERE code = 'about_stuff_city' ;
Delete rows using non-unique columns
Deleting rows using non-unique columns removes every row that returns TRUE
for the WHERE
clause's a_expr
. This can easily result in deleting data you didn't intend to.
> DELETE FROM promo_codes WHERE creation_time > '2019-01-30 00:00:00+00:00' ;
The example statement deleted four rows, which might be unexpected.
Return deleted rows
To see which rows your statement deleted, include the RETURNING
clause to retrieve them using the columns you specify.
Use all columns
By specifying *
, you retrieve all columns of the delete rows.
Use specific columns
To retrieve specific columns, name them in the RETURNING
clause.
> DELETE FROM promo_codes WHERE creation_time > '2019-01-29 00:00:00+00:00' RETURNING code , rules ;
code | rules +------------------------+----------------------------------------------+ box_investment_stuff | {"type": "percent_discount", "value": "10%"} energy_newspaper_field | {"type": "percent_discount", "value": "10%"} simple_guy_theory | {"type": "percent_discount", "value": "10%"} study_piece_war | {"type": "percent_discount", "value": "10%"} tv_this_list | {"type": "percent_discount", "value": "10%"} (5 rows)
Change column labels
When RETURNING
specific columns, you can change their labels using AS
.
> DELETE FROM promo_codes WHERE creation_time > '2019-01-28 00:00:00+00:00' RETURNING code , rules AS discount ;
code | discount +---------------------+----------------------------------------------+ chair_company_state | {"type": "percent_discount", "value": "10%"} view_reveal_radio | {"type": "percent_discount", "value": "10%"} (2 rows)
Sort and return deleted rows
To sort and return deleted rows, use a statement like the following:
> WITH a AS ( DELETE FROM promo_codes WHERE creation_time > '2019-01-27 00:00:00+00:00' RETURNING * ) SELECT * FROM a ORDER BY expiration_time ;
code | description | creation_time | expiration_time | rules +----------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+---------------------------+---------------------------+----------------------------------------------+ often_thing_hair | Society right wish face see if pull. Great generation social bar read budget wonder natural. Somebody dark field economic material. Nature nature paper law worry common. Serious activity hospital wide none. | 2019-01-27 03:04:05+00:00 | 2019-01-29 03:04:05+00:00 | {"type": "percent_discount", "value": "10%"} step_though_military | Director middle summer most create any. | 2019-01-27 03:04:05+00:00 | 2019-01-29 03:04:05+00:00 | {"type": "percent_discount", "value": "10%"} own_whose_economy | Social participant order this. Guy toward nor indeed police player inside nor. Model education voice several college art on. Start listen their maybe. | 2019-01-27 03:04:05+00:00 | 2019-01-30 03:04:05+00:00 | {"type": "percent_discount", "value": "10%"} crime_experience_certainly | Prepare right teacher mouth student. Trouble condition weight during scene something stand. | 2019-01-27 03:04:05+00:00 | 2019-01-31 03:04:05+00:00 | {"type": "percent_discount", "value": "10%"} policy_its_wife | Player either she something good minute or. Nearly policy player receive. Somebody mean book store fire realize. | 2019-01-27 03:04:05+00:00 | 2019-01-31 03:04:05+00:00 | {"type": "percent_discount", "value": "10%"} (5 rows)
Delete with index hints
Suppose you create a multi-column index on the users
table with the name
and city
columns.
> CREATE INDEX ON users ( name , city );
Now suppose you want to delete the two users named "Jon Snow". You can use the EXPLAIN (OPT)
command to see how the cost-based optimizer decides to perform the delete:
> EXPLAIN ( OPT ) DELETE FROM users WHERE name = 'Jon Snow' ;
text ------------------------------------------------------------------------------------ delete users ├── scan users@users_name_city_idx │ └── constraint: /10/9/8: [/'Jon Snow' - /'Jon Snow'] └── f-k-checks ├── f-k-checks-item: vehicles(city,owner_id) -> users(city,id) │ └── semi-join (lookup vehicles@vehicles_auto_index_fk_city_ref_users) │ ├── with-scan &1 │ └── filters (true) ├── f-k-checks-item: rides(city,rider_id) -> users(city,id) │ └── semi-join (lookup rides@rides_auto_index_fk_city_ref_users) │ ├── with-scan &1 │ └── filters (true) └── f-k-checks-item: user_promo_codes(city,user_id) -> users(city,id) └── semi-join (lookup user_promo_codes) ├── with-scan &1 └── filters (true) (16 rows)
The output of the EXPLAIN
statement shows that the optimizer scans the newly-created users_name_city_idx
index when performing the delete. This makes sense, as you are performing a delete based on the name
column.
Now suppose that instead you want to perform a delete, but using the id
column instead.
> EXPLAIN ( OPT ) DELETE FROM users WHERE id IN ( '70a3d70a-3d70-4400-8000-000000000016' , '3d70a3d7-0a3d-4000-8000-00000000000c' );
text --------------------------------------------------------------------------------------------------------------- delete users ├── select │ ├── scan users@users_name_city_idx │ └── filters │ └── users.id IN ('3d70a3d7-0a3d-4000-8000-00000000000c', '70a3d70a-3d70-4400-8000-000000000016') └── f-k-checks ├── f-k-checks-item: vehicles(city,owner_id) -> users(city,id) │ └── semi-join (hash) │ ├── with-scan &1 │ ├── scan vehicles@vehicles_auto_index_fk_city_ref_users │ └── filters │ ├── city = vehicles.city │ └── id = owner_id ├── f-k-checks-item: rides(city,rider_id) -> users(city,id) │ └── semi-join (lookup rides@rides_auto_index_fk_city_ref_users) │ ├── with-scan &1 │ └── filters (true) └── f-k-checks-item: user_promo_codes(city,user_id) -> users(city,id) └── semi-join (hash) ├── with-scan &1 ├── scan user_promo_codes └── filters ├── city = user_promo_codes.city └── id = user_id (24 rows)
The optimizer still scans the newly-created users_name_city_idx
index when performing the delete. Although scanning the table on this index could still be the most efficient, you may want to assess the performance difference between using users_name_city_idx
and an index on the id
column, as you are performing a delete with a filter on the id
column.
If you provide an index hint (i.e., force the index selection) to use the primary index on the column instead, the CockroachDB will scan the users table using the primary index, on city
, and id
.
> EXPLAIN ( OPT ) DELETE FROM users @ users_pkey WHERE id IN ( '70a3d70a-3d70-4400-8000-000000000016' , '3d70a3d7-0a3d-4000-8000-00000000000c' );
text --------------------------------------------------------------------------------------------------------------- delete users ├── select │ ├── scan users │ │ └── flags: force-index=primary │ └── filters │ └── users.id IN ('3d70a3d7-0a3d-4000-8000-00000000000c', '70a3d70a-3d70-4400-8000-000000000016') └── f-k-checks ├── f-k-checks-item: vehicles(city,owner_id) -> users(city,id) │ └── semi-join (hash) │ ├── with-scan &1 │ ├── scan vehicles@vehicles_auto_index_fk_city_ref_users │ └── filters │ ├── city = vehicles.city │ └── id = owner_id ├── f-k-checks-item: rides(city,rider_id) -> users(city,id) │ └── semi-join (lookup rides@rides_auto_index_fk_city_ref_users) │ ├── with-scan &1 │ └── filters (true) └── f-k-checks-item: user_promo_codes(city,user_id) -> users(city,id) └── semi-join (hash) ├── with-scan &1 ├── scan user_promo_codes └── filters ├── city = user_promo_codes.city └── id = user_id (25 rows)
See also
-
INSERT
-
UPDATE
-
UPSERT
-
TRUNCATE
-
ALTER TABLE
-
DROP TABLE
-
DROP DATABASE
- SQL Statements
- Limit Query Results
- Delete Data
- Bulk-delete data
- Batch Delete Expired Data with Row-Level TTL
Was this helpful?
Source: https://www.cockroachlabs.com/docs/stable/delete.html
0 Response to "Primary Key Continue When Row Deleted"
Post a Comment