官术网_书友最值得收藏!

Storing data using AWS S3

There are many cases where we just want to save content that we scrape into a local copy for archive purposes, backup, or later bulk analysis. We also might want to save media from those sites for later use. I've built scrapers for advertisement compliance companies, where we would track and download advertisement based media on web sites to ensure proper usage, and also to store for later analysis, compliance and transcoding.

The storage required for these types of systems can be immense, but with the advent of cloud storage services such as AWS S3 (Simple Storage Service), this becomes much easier and more cost effective than managing a large SAN (Storage Area Network) in your own IT department. Plus, S3 can also automatically move data from hot to cold storage, and then to long-term storage, such as a glacier, which can save you much more money.

We won't get into all of those details, but simply look at storing our planets.html file into an S3 bucket. Once you can do this, you can save any content you want to year hearts desire.

主站蜘蛛池模板: 青州市| 扬州市| 揭阳市| 左贡县| 清新县| 铅山县| 扬中市| 丹江口市| 南开区| 天等县| 松溪县| 盐亭县| 封丘县| 天长市| 黄陵县| 西乡县| 正蓝旗| 平原县| 元阳县| 普陀区| 沁阳市| 仙游县| 梅州市| 广汉市| 赤壁市| 陆河县| 西充县| 永泰县| 兴隆县| 彰化市| 弋阳县| 大连市| 鞍山市| 北川| 八宿县| 宣恩县| 额济纳旗| 伊金霍洛旗| 佛学| 望都县| 巍山|