MySQL基础知识:创建MySQL数据库和表
2021/6/10 19:28:37
本文主要是介绍MySQL基础知识:创建MySQL数据库和表,对大家解决编程问题具有一定的参考价值,需要的程序猿们随着小编来一起学习吧!
虚构一个微型在线书店的数据库和数据,作为后续MySQL脚本的执行源,方便后续MySQL和SQL的练习。
在虚构这个库的过程中,主要涉及的是如何使用命令行管理 MySQL数据库对象:数据库、表、索引、外键等;另一个更为重要的是如何Mock对应表的数据。
虚构书店数据库的dump脚本:Github
数据库(Database)将要创建的虚拟书店的数据库名为: mysql_practice;
创建数据库的语法:
CREATE DATABASE [IF NOT EXISTS] database_name [CHARACTER SET charset_name] [COLLATE collation_name]
- IF NOT EXISTS: 可选项,避免数据库已经存在时报错。
- CHARACTER SET:可选项,不指定的时候会默认给个。查看当前MySQL Server支持的字符集(character set):show character set; -- 方法1 show charset; -- 方法2 show char set; -- 方法3
- COLLATE:针对特定character set比较字符串的规则集合;可选项,不指定的时候会默认给个。获取 charater set的 collationsshow collation like 'utf8%'; collation名字的规则: charater_set_name_ci 或者 charater_set_name_cs 或 charater_set_name_bin;_ci表示不区分大小写,_cs表示区分大小写;_bin表示用编码值比较。
- 示例:CREATE DATABASE my_test_tb CHARACTER SET utf8mb4 COLLATE utf8mb4_0900_ai_ci;
TODO: 关于 character set和collations,内容稍微有点多,后面会单独记一篇文章。
登录的时候选择数据库mysql -uroot -D database_name -p登录后选择数据库
use database_name;查看当前选的数据库
select database();创建新数据库
create database if not exists mysql_practice;
通过下面的语句可以检查创建的数据库:
show create database mysql_practice;
可以看到,如果创建数据库时候没有指定 character set 和 collate 的话,会默认指定一套。
显示所有当前账户可见的数据库show databases;删除数据库
drop database if exists mysql_practice;
MySQL中 schema 是 database 的同义词,因此也可以使用下面语句删除数据库:
drop schema if exists mysql_practice;数据表(Table) MySQL创建数据表的语法
CREATE TABLE [IF NOT EXISTS] table_name( column_1_definition, column_2_definition, ..., table_constraints ) ENGINE=storage_engine;
表列的定义语法:
column_name data_type(length) [NOT NULL] [DEFAULT value] [AUTO_INCREMENT] column_constraint;
表的约束(Table Constraints): UNIQUE, CHECK, PRIMARY KEY and FOREIGN KEY.
查看表的定义desc table_name;创建mysql_practice数据表
USE mysql_practice; DROP TABLE IF EXISTS customer_order; DROP TABLE IF EXISTS book; DROP TABLE IF EXISTS book_category; DROP TABLE IF EXISTS customer_address; DROP TABLE IF EXISTS customer; DROP TABLE IF EXISTS region; -- region,数据使用: https://github.com/xiangyuecn/AreaCity-JsSpider-StatsGov CREATE TABLE IF NOT EXISTS region( id INT AUTO_INCREMENT, pid INT NOT NULL, deep INT NOT NULL, name VARCHAR(200) NOT NULL, pinyin_prefix VARCHAR(10) NOT NULL, pinyin VARCHAR(200) NOT NULL, ext_id VARCHAR(100) NOT NULL, ext_name VARCHAR(200) NOT NULL, PRIMARY KEY(id) ); -- customer CREATE TABLE IF NOT EXISTS customer( id INT AUTO_INCREMENT, no VARCHAR(50) NOT NULL, first_name VARCHAR(255) NOT NULL, last_name VARCHAR(255) NOT NULL, status VARCHAR(20) NOT NULL, phone_number VARCHAR(20) NULL, updated_at DATETIME NOT NULL, created_at DATETIME NOT NULL, PRIMARY KEY(id), unique(no) ) ENGINE=INNODB; -- customer address CREATE TABLE IF NOT EXISTS customer_address( id INT AUTO_INCREMENT, customer_id INT NOT NULL, area_id INT NULL, address_detail VARCHAR(200) NULL, is_default bit NOT NULL, updated_at DATETIME NOT NULL, created_at DATETIME NOT NULL, PRIMARY KEY(id), FOREIGN KEY(customer_id) REFERENCES customer (id) ON UPDATE RESTRICT ON DELETE CASCADE ) ENGINE=INNODB; -- book category CREATE TABLE IF NOT EXISTS book_category( id INT AUTO_INCREMENT, code VARCHAR(200) NOT NULL, name VARCHAR(200) NOT NULL, parent_id INT NULL, deep INT NULL, updated_at DATETIME NOT NULL, created_at DATETIME NOT NULL, PRIMARY KEY(id) ); -- book CREATE TABLE IF NOT EXISTS book( id INT AUTO_INCREMENT, category_id INT NOT NULL, no VARCHAR(50) NOT NULL, name VARCHAR(200) NOT NULL, status VARCHAR(50) NOT NULL, unit_price DOUBLE NOT NULL, author VARCHAR(50) NULL, publish_date DATETIME NULL, publisher VARCHAR(200) NOT NULL, updated_at DATETIME NOT NULL, created_at DATETIME NOT NULL, PRIMARY KEY(id), FOREIGN KEY (category_id) REFERENCES book_category (id) ON UPDATE RESTRICT ON DELETE CASCADE ); -- orders CREATE TABLE IF NOT EXISTS customer_order( id INT AUTO_INCREMENT, no VARCHAR(50) NOT NULL, customer_id INT NOT NULL, book_id INT NOT NULL, quantity INT NOT NULL, total_price DOUBLE NOT NULL, discount DOUBLE NULL, order_date DATETIME NOT NULL, updated_at DATETIME NOT NULL, created_at DATETIME NOT NULL, PRIMARY KEY(id), FOREIGN KEY (customer_id) REFERENCES customer(id) ON UPDATE RESTRICT ON DELETE CASCADE, FOREIGN KEY (book_id) references book (id) on update restrict on delete cascade ) ENGINE=INNODB;导入region数据
下载region csv数据:【三级】省市区 数据下载.
导入语句:
LOAD DATA INFILE '/tmp/ok_data_level3.csv' INTO TABLE region FIELDS TERMINATED BY ',' ENCLOSED BY '"' LINES TERMINATED BY '\n' IGNORE 1 ROWS;
导入如果报错:
ERROR 1290 (HY000): The MySQL server is running with the --secure-file-priv option so it cannot execute this statement
- 通过命令 mdfind -name my.cnf 找到mysql配置文件 my.cnf;
- 解决办法 (还没实际测试过,大都使用的是 LOATA DATA LOCAL INFILE 方式)
或者使用 LOAD DATA LOCAL INFILE代替 LOAD DATA INFILE 即:
LOAD DATA LOCAL INFILE '/tmp/ok_data_level3.csv' INTO TABLE region FIELDS TERMINATED BY ',' ENCLOSED BY '"' LINES TERMINATED BY '\n' IGNORE 1 ROWS;
如果报错:
Error Code: 3948. Loading local data is disabled; this must be enabled on both the client and server sides
或者报错:
ERROR 1148 (42000): The used command is not allowed with this MySQL version
- 查看配置: show variables like "local_infile";
- 修改配置: set global local_infile = 1;
创建一个SP:
USE mysql_practice; DROP PROCEDURE IF EXISTS sp_generate_customers; DELIMITER $$ CREATE PROCEDURE sp_generate_customers() BEGIN -- Generate 10000 customer and customer_address set @fNameIndex = 1; set @lNameIndex = 1; loop_label_f: LOOP IF @fNameIndex > 100 THEN LEAVE loop_label_f; END IF; set @fName = ELT(@fNameIndex, "James","Mary","John","Patricia","Robert","Linda","Michael","Barbara","William","Elizabeth","David","Jennifer","Richard","Maria","Charles","Susan","Joseph","Margaret","Thomas","Dorothy","Christopher","Lisa","Daniel","Nancy","Paul","Karen","Mark","Betty","Donald","Helen","George","Sandra","Kenneth","Donna","Steven","Carol","Edward","Ruth","Brian","Sharon","Ronald","Michelle","Anthony","Laura","Kevin","Sarah","Jason","Kimberly","Matthew","Deborah","Gary","Jessica","Timothy","Shirley","Jose","Cynthia","Larry","Angela","Jeffrey","Melissa","Frank","Brenda","Scott","Amy","Eric","Anna","Stephen","Rebecca","Andrew","Virginia","Raymond","Kathleen","Gregory","Pamela","Joshua","Martha","Jerry","Debra","Dennis","Amanda","Walter","Stephanie","Patrick","Carolyn","Peter","Christine","Harold","Marie","Douglas","Janet","Henry","Catherine","Carl","Frances","Arthur","Ann","Ryan","Joyce","Roger","Diane"); loop_label_last: LOOP IF @lNameIndex > 100 THEN LEAVE loop_label_last; END IF; SET @lName = ELT(@lNameIndex, "Smith","Johnson","Williams","Jones","Brown","Davis","Miller","Wilson","Moore","Taylor","Anderson","Thomas","Jackson","White","Harris","Martin","Thompson","Garcia","Martinez","Robinson","Clark","Rodriguez","Lewis","Lee","Walker","Hall","Allen","Young","Hernandez","King","Wright","Lopez","Hill","Scott","Green","Adams","Baker","Gonzalez","Nelson","Carter","Mitchell","Perez","Roberts","Turner","Phillips","Campbell","Parker","Evans","Edwards","Collins","Stewart","Sanchez","Morris","Rogers","Reed","Cook","Morgan","Bell","Murphy","Bailey","Rivera","Cooper","Richardson","Cox","Howard","Ward","Torres","Peterson","Gray","Ramirez","James","Watson","Brooks","Kelly","Sanders","Price","Bennett","Wood","Barnes","Ross","Henderson","Coleman","Jenkins","Perry","Powell","Long","Patterson","Hughes","Flores","Washington","Butler","Simmons","Foster","Gonzales","Bryant","Alexander","Russell","Griffin","Diaz","Hayes"); -- insert into customer INSERT INTO customer(no, first_name, last_name, status, phone_number, updated_at, created_at) values( REPLACE(LEFT(uuid(), 16), '-', ''), @fName, @lName, 'ACTIVE', null, curdate(), curdate() ); -- insert into customer_address set @randomArea = 0; SELECT id into @randomArea FROM region where deep = 2 ORDER BY RAND() LIMIT 1; INSERT INTO customer_address(customer_id, area_id, address_detail, is_default, updated_at, created_at) VALUES( @@Identity, @randomArea, '', 1, curdate(), curdate() ); set @lNameIndex = @lNameIndex + 1; END LOOP loop_label_last; SET @lNameIndex = 1; -- Note: assign 1 to last name index, for next loop. SET @fnameIndex = @fnameIndex + 1; END LOOP loop_label_f; -- update address_detail in customer_address UPDATE customer_address ca JOIN region r on ca.area_id = r.id and r.deep = 2 join region r2 on r.pid = r2.id and r2.deep = 1 join region r3 on r2.pid = r3.id and r3.deep = 0 SET ca.address_detail = concat(r3.ext_name, r2.ext_name, r.ext_name); END $$ DELIMITER ;
调用SP:
call sp_generate_customers();生成产品分类和产品数据
第零步: 手动插入产品分类到product_category表中
INSERT INTO product_category(code,name, parent_id, deep, updated_at, created_at) VALUES ('BOOK', 'Book', 0, 0, curdate(), curdate()), ('BOOK_CODE', 'Code Book', 1, 1, curdate(), curdate()), ('BOOK_CHIDREN', 'Children Book', 1, 1, curdate(), curdate()), ('BOOK_SCIENCE', 'Science Book', 1, 1, curdate(), curdate());
第一步: 用Python写个爬虫工具,抓取书店的商品信息。
下面是抓取当当搜索“科学”关键字的书籍列表。
import requests import csv from bs4 import BeautifulSoup def crawl(url): res = requests.get(url) res.encoding = 'gb18030' soup = BeautifulSoup(res.text, 'html.parser') n = 0 section = soup.find('ul', id='component_59') allLIs = section.find_all('li') #print(allLIs) with open('output_science.csv', 'w', encoding='utf8') as f: csv_writer = csv.writer(f, delimiter='#') # 由于内容里有',',因此这里指定'#'作为csv分隔符 csv_writer.writerow(['序号', '书名', '价格', '作者', '出版时间', '出版社']) for books in allLIs: title = books.select('.name')[0].text.strip().split(' ', 1)[0].strip() price = books.select('.search_pre_price')[0].text.strip('¥') authorInfo = books.select('.search_book_author')[0].text.strip().split('/') author = authorInfo[0] publishDate = authorInfo[1] publisher = authorInfo[2] n += 1 csv_writer.writerow([n, title, price, author, publishDate, publisher]) url = 'http://search.dangdang.com/?key=%BF%C6%D1%A7&act=input' crawl(url)
第二步: 导入csv数据到MySQL数据表mock_science中。
CREATE TABLE `mock_science` ( `id` int(11) NOT NULL, `name` varchar(200) DEFAULT NULL, `price` double DEFAULT NULL, `author` varchar(100) DEFAULT NULL, `publish_date` varchar(100) DEFAULT NULL, `publisher` varchar(100) DEFAULT NULL, PRIMARY KEY (`id`) ) ENGINE=InnoDB DEFAULT CHARSET=utf8mb4 COLLATE=utf8mb4_0900_ai_ci;
第三步: 插入科学类书信息到product表中
INSERT book(category_id, no, name, status,unit_price, author,publish_date,publisher, updated_at, created_at) SELECT 4, REPLACE(LEFT(uuid(), 16), '-', ''), name, 'ACTIVE', price, author, publish_date, publisher, curdate(), curdate() FROM mock_science;
循环第一到第三步,可以插入更多的产品信息。练习数据库最终抓取了JAVA,儿童,科学三个关键搜索出的第一页书籍。
生成订单数据随机生成订单数据的SP(注意:这个sp生成的数据,还需要进一步处理):
USE mysql_practice; DROP PROCEDURE IF EXISTS sp_generate_orders; DELIMITER $$ -- Reference: https://www.mysqltutorial.org/select-random-records-database-table.aspx -- Generate orders for last two years. -- each day have orders range: [500, 5000] CREATE PROCEDURE sp_generate_orders() BEGIN SET @startDate = '2020-03-01'; SET @endDate = curdate(); loop_label_p: LOOP IF @startDate > @endDate THEN LEAVE loop_label_p; END IF; SET @randCustomerTotal = FLOOR(RAND()*50) + 100; SET @randBookTotal = FLOOR(RAND()*5) + 1; SET @randQty = FLOOR(RAND()*3) + 1; SET @query1 = CONCAT('INSERT INTO customer_order(no, customer_id, book_id, quantity, total_price,discount, order_date, updated_at, created_at)'); SET @query1 = CONCAT(@query1, ' select ', "'", uuid(), "'",', c.id, p.id,', @randQty, ', 0, 0, ', "'",@startDate,"'", ',', "'",curdate(),"'" ,',', "'",curdate(),"'"); SET @query1 = CONCAT(@query1, ' FROM (select id from customer ORDER BY RAND() LIMIT ', @randCustomerTotal,') c join '); SET @query1 = CONCAT(@query1, ' (select id from book order by rand() limit ', @randBookTotal,') p '); SET @query1 = CONCAT(@query1, 'where c.id is not null'); PREPARE increased FROM @query1; EXECUTE increased; SET @startDate = DATE_ADD(@startDate, INTERVAL 1 DAY); END LOOP loop_label_p; END $$ DELIMITER ;
总共会生成几十万或上百万条order数据;最好先简单加下index,不然query太慢,可以在创建db table后就加上。
添加index:
ALTER TABLE book ADD INDEX idx_unit_price(unit_price); ALTER TABLE customer_order ADD INDEX idx_order_no(no); ALTER TABLE customer_order ADD INDEX idx_order_date(order_date); ALTER TABLE customer_order ADD INDEX idx_quantity(quantity);
更新order no:
-- update order total_price -- please note it is better to add index first. otherwise it will be slow. -- update order_no update customer_order set no = concat(REPLACE(LEFT(no, 16), '-', ''), customer_id, book_id) where no is not null; -- update total price
如果不想有重复的order no,可以通过下面的sql更新order no:
-- 处理重复的 order no update customer_order co join (select no from customer_order co2 group by co2.no having count(*) > 1) as cdo on co.no = cdo.no set co.no = concat(REPLACE(LEFT(uuid(), 16), '-', ''), customer_id, book_id);
如果还有重复的order no,继续run上面这个sql,直到没有重复的即可。
更新order表里的total_price:
-- update total price update customer_order co join book b on co.book_id = b.id SET co.total_price = co.quantity * b.unit_price;
至此,我们的数据库表和对应的mock数据已经基本完成。使用mysqldump备份一下:
mysqldump -u [username] –p[password] [database_name] > [dump_file.sql]下一步
- 视图(View)
- 存储过程(Store Procedure)
- 函数(Function)
- 触发器(Trigger)
- 定时任务(Job)
原文链接:
https://www.cnblogs.com/codesee/p/14510886.html
如果觉得本文对你有帮助,可以关注一下我公众号,回复关键字【面试】即可得到一份Java核心知识点整理与一份面试大礼包!另有更多技术干货文章以及相关资料共享,大家一起学习进步!
这篇关于MySQL基础知识:创建MySQL数据库和表的文章就介绍到这儿,希望我们推荐的文章对大家有所帮助,也希望大家多多支持为之网!
- 2024-11-20部署MySQL集群入门:新手必读指南
- 2024-11-20部署MySQL集群教程:初学者指南
- 2024-11-20部署MySQL集群项目实战:新手教程
- 2024-11-20部署MySQL集群资料:新手入门教程
- 2024-11-20MySQL集群部署教程:入门级详解
- 2024-11-20MySQL集群教程:入门与实践指南
- 2024-11-20部署MySQL集群教程:新手入门指南
- 2024-11-20MySQL读写分离教程:轻松入门
- 2024-11-20部署MySQL集群入门:一步一步搭建你的数据库集群
- 2024-11-19部署MySQL集群学习:入门教程