Lab_3_SysOps_Storage_Linux_v2.5
System Operations - Lab 3: Managing Storage in AWS (Linux) - 2.5
================================================================================================================== Using this command reference. ================================================================================================================== 1. Locate the section you need. Each section in this file matches a section in the lab instructions. 2. Replace items in angle brackets - < > - with appropriate values. For example, in this command you would replace the value - <JobFlowID> - (including the brackets) with the parameter indicated in the lab instructions: elastic-mapreduce --list <JobFlowID>. You can also use find and replace to change bracketed parameters in bulk. 3. Do NOT enable the Word Wrap feature in Windows Notepad or the text editor you use to view this file. ++++1. Configuring a New Instance++++ ==================================================================================================================
1.5 Create Amazon EC2 Instance
================================================================================================================== 1.5.8 Paste text into User data field #!/bin/bash
yum update -y ++++2. Taking Snapshots of Your Instance++++ ==================================================================================================================
2.1 Take an Initial Snapshot
================================================================================================================== 2.1.5 Obtain a full description of the Processor instance aws ec2 describe-instances --filter 'Name=tag:Name,Values=Processor' 2.1.6 Refine description to include only VolumeId of Amazon EBS volume aws ec2 describe-instances --filter 'Name=tag:Name,Values=Processor' --query 'Reservations[0].Instances[0].BlockDeviceMappings[0].Ebs.{VolumeId:VolumeId}' 2.1.7 Obtain instance ID of Processor instance aws ec2 describe-instances --filters 'Name=tag:Name,Values=Processor' --query 'Reservations[0].Instances[0].InstanceId' 2.1.8 Stop Processor instance aws ec2 stop-instances --instance-ids <instance-id> 2.1.9 Check status of stop action aws ec2 describe-instance-status --instance-id <instance-id> 2.1.10 Create first snapshot aws ec2 create-snapshot --volume-id <volume-id> 2.1.11 Report on snapshot status aws ec2 describe-snapshots --snapshot-id <snapshot-id> 2.1.12 Restart Processor instance aws ec2 start-instances --instance-ids <instance-id> 2.1.13 Check status of start action aws ec2 describe-instance-status --instance-id <instance-id> ==================================================================================================================
2.2 Schedule Creation of Subsequent Snapshots
================================================================================================================== 2.2.1 Create the cron task description echo "* * * * * aws ec2 create-snapshot --volume-id <volume-id> 2>&1 >> /tmp/cronlog" > cronjob 2.2.2 Schedule the cron task crontab cronjob 2.2.3 Verify that snapshots are being created aws ec2 describe-snapshots --filters "Name=volume-id,Values=<volume-id>" ==================================================================================================================
2.3 Retain Only Last Two Snapshots
================================================================================================================== 2.3.1 Remove cron entry crontab -r 2.3.2 Examine snapshotter.py more snapshotter.py 2.3.3 Verify the current number of snapshots for the volume aws ec2 describe-snapshots --filters "Name=volume-id,Values=<volume-id>" --query 'Snapshots[*].SnapshotId' 2.3.4 Execute snapshotter.py python snapshotter.py 2.3.5 Re-examine the current number of snapshots for the volume aws ec2 describe-snapshots --filters "Name=volume-id,Values=<volume-id>" --query 'Snapshots[*].SnapshotId' ++++3. Moving Log Files to Amazon S3++++ ==================================================================================================================
3.2 Move a Log File into Amazon S3
================================================================================================================== 3.2.1 Download loggen.sh wget https://d2lrzjb0vjvpn5.cloudfront.net/sys-ops/v2.5/lab-3-storage-linux/static/loggen.sh 3.2.2 Change permission on loggen.sh chmod 740 loggen.sh 3.2.3 Run loggen.sh in the background ./loggen.sh & 3.2.4 Verify end of timestamp file tail timestamp.log 3.2.5 Move file into Amazon S3 aws s3 mv timestamp.log s3://<s3-bucket-name>/logfiles/timestamp-`date +"%m-%d-%Y-%H.%M.%S"`.log 3.2.6 Move file again aws s3 mv timestamp.log s3://<s3-bucket-name>/logfiles/timestamp-`date +"%m-%d-%Y-%H.%M.%S"`.log 3.2.7 List the contents of your Amazon S3 bucket aws s3 ls s3://<s3-bucket-name> 3.2.8 List all objects containing the /logfiles/ prefix aws s3 ls s3://<s3-bucket-name>/logfiles/ 3.2.9 Move oldest file to /logfiles/archive/ prefix aws s3 mv s3://<s3-bucket-name>/logfiles/<file-name> s3://<s3-bucket-name>/logfiles/archive/<file-name> 3.2.10 Verify move aws s3 ls s3://<s3-bucket-name>/logfiles/ ++++4. Challenge Solution - Synchronize Folder with Amazon S3++++ ==================================================================================================================
4.1 Download and Unzip Sample Files
================================================================================================================== 4.1.1 Download files wget https://d2lrzjb0vjvpn5.cloudfront.net/sys-ops/v2.5/lab-3-storage-linux/static/files.zip 4.1.2 Unzip the directory unzip files.zip ==================================================================================================================
4.2 Synchronize Files
================================================================================================================== 4.2.1 Set versioning on bucket aws s3api put-bucket-versioning --bucket <s3-bucket-name> --versioning-configuration Status=Enabled 4.2.2 Synchronize files aws s3 sync files s3://<s3-bucket-name>/files/ 4.2.3 Confirm your files on Amazon S3 aws s3 ls s3://<s3-bucket-name>/files/ 4.2.4 Delete file1.txt rm files/file1.txt 4.2.5 Synchronize, requesting deletion of files no longer in the source directory aws s3 sync files s3://<s3-bucket-name>/files/ --delete 4.2.6 Verify that the item was deleted in Amazon S3 aws s3 ls s3://<s3-bucket-name>/files/ 4.2.7 Find past versions of file1.txt aws s3api list-object-versions --bucket <s3-bucket-name> --prefix files/file1.txt 4.2.8 Restore previous version of file to local file system aws s3api get-object --bucket <s3-bucket-name> --key files/file1.txt --version-id <version-id> files/file1.txt 4.2.9 Verify that the file has been restored locally ls files 4.2.10 Re-sync the folder to Amazon S3 aws s3 sync files s3://<s3-bucket-name>/files/ 4.2.11 Verify that the file has been restored to the bucket aws s3 ls s3://<s3-bucket-name>/files/ 漏 2016 Amazon Web Services, Inc. or its affiliates. All rights reserved.
Lab_3_SysOps_Storage_Linux_v2.5的更多相关文章
随机推荐
- Maven构件解析(转)
文章转自http://gavinwind2000.iteye.com/blog/2290652 谢谢博主的总结! 在Maven中,任何一个依赖.插件或者项目构建的输出,都可以称之为构件. Maven在 ...
- 实验一、熟悉DOS命令
实验一.熟悉DOS命令 实验一.熟悉DOS命令 一. 实验目的 1.从操作系统理论的观点来了解和掌握DOS有关用户接口的特点: 2.熟悉常用DOS操作命令(md.cd.rd.copy.m ...
- re.S
在Python的正则表达式中,有一个参数为re.S.它表示多行匹配
- c++ 顺序容器适配器
第一次看到stack,以为它是一个和vector同等地位的容器,其实不是 官方解释:stacks are a type of container adaptor, specifically desig ...
- linux进程命令
1. ps命令 命令 命令名称: 命令名称: ps 功能: 功能: 查询正在执行的进程 语法: 可选参数] 语法: ps [可选参数] 描述: 命令提供 命令提供Linux系统中正在发生的事情的 ...
- Unity3D-坐标转换笔记
Transform.TransformPoint 作用 : 将一个点从以自身为坐标系的本地坐标转换成世界坐标 Transform.InverseTransformPoint 作用 : 将一个点从世界坐 ...
- xpath轴的正确使用姿势
网上看了许多关于轴的介绍,只介绍了语法,而没有明说具体实际中该怎么使用,百思不得其解. 背景--python中使用xpath: ----------------------------------- ...
- hhvm之轻进程
本文为原创,转载请注明:http://www.cnblogs.com/gistao/ 背景 我们在aws上部署了hhvm,高峰段发现cpu idle降的比较低,只有10-20%,而使用php-fpm的 ...
- 两个list取不同值
转自同名博文,未知真正出处,望作者见谅 如题:有List<String> list1和List<String> list2,两个集合各有上万个元素,怎样取出两个集合中不同的元素 ...
- HDU 4857 逃生 (反向拓扑排序 & 容器实现)
题目链接:http://acm.hdu.edu.cn/showproblem.php?pid=4857 逃生 Time Limit: 2000/1000 MS (Java/Others) Mem ...