Rename License file and add Contributing sections

* Make it clear from project root we are using MIT-LICENSE
* Add contributing section w/ overview of how to submit code to repo
This commit is contained in:
Harlow Ward 2015-05-23 08:52:08 -07:00
parent bde3e96ad9
commit 9077d288d2
3 changed files with 34 additions and 5 deletions

13
CONTRIBUTING.md Normal file
View file

@ -0,0 +1,13 @@
# Contributing to Golang Kinesis Connectors
Pull requests are much apprecited. Please help make the project a success!
To contribute:
1. Fork the [official repository][1].
2. Make your changes in a topic branch.
3. Squash commits and add an [excellent commit message][2].
4. Send a pull request.
[1]: https://github.com/harlow/kinesis-connectors/tree/master
[2]: http://tbaggery.com/2008/04/19/a-note-about-git-commit-messages.html

View file

@ -1,12 +1,15 @@
# Golang Kinesis Connectors # Golang Kinesis Connectors
__Note:__ _This codebase is a under active development. Expect breaking changes until 1.0 version release._ __Note:__ _This codebase is a under active development._
### Kinesis connector applications written in Go ### Kinesis connector applications written in Go
Inspired by the [Amazon Kinesis Connector Library][1]. These components are used for extracting streaming event data Inspired by the [Amazon Kinesis Connector Library][1]. These components are used for extracting streaming event data
into S3, Redshift, DynamoDB, and more. See the [API Docs][2] for package documentation. into S3, Redshift, DynamoDB, and more. See the [API Docs][2] for package documentation.
[1]: https://github.com/awslabs/amazon-kinesis-connectors
[2]: http://godoc.org/github.com/harlow/kinesis-connectors
## Overview ## Overview
Each Amazon Kinesis connector application is a pipeline that determines how records from an Amazon Kinesis stream will be handled. Records are retrieved from the stream, transformed according to a user-defined data model, buffered for batch processing, and then emitted to the appropriate AWS service. Each Amazon Kinesis connector application is a pipeline that determines how records from an Amazon Kinesis stream will be handled. Records are retrieved from the stream, transformed according to a user-defined data model, buffered for batch processing, and then emitted to the appropriate AWS service.
@ -35,7 +38,9 @@ The S3 Pipeline performs the following steps:
2. Upload the records to an S3 bucket. 2. Upload the records to an S3 bucket.
3. Set the current Shard checkpoint in Redis. 3. Set the current Shard checkpoint in Redis.
The config vars are loaded done with [gcfg][3]. The config vars are loaded done with [gcfg].
[gcfg]: https://code.google.com/p/gcfg/
```go ```go
package main package main
@ -118,6 +123,17 @@ func main() {
} }
``` ```
[1]: https://github.com/awslabs/amazon-kinesis-connectors ## Contributing
[2]: http://godoc.org/github.com/harlow/kinesis-connectors
[3]: https://code.google.com/p/gcfg/ Please see [CONTRIBUTING.md].
Thank you, [contributors]!
[LICENSE]: /MIT-LICENSE
[CONTRIBUTING.md]: /CONTRIBUTING.md
## License
Copyright (c) 2015 Harlow Ward. It is free software, and may
be redistributed under the terms specified in the [LICENSE] file.
[contributors]: https://github.com/harlow/kinesis-connectors/graphs/contributors