Adoptable Cookbooks List

Looking for a cookbook to adopt? You can now see a list of cookbooks available for adoption!
List of Adoptable Cookbooks

Supermarket Belongs to the Community

Supermarket belongs to the community. While Chef has the responsibility to keep it running and be stewards of its functionality, what it does and how it works is driven by the community. The chef/supermarket repository will continue to be where development of the Supermarket application takes place. Come be part of shaping the direction of Supermarket by opening issues and pull requests or by joining us on the Chef Mailing List.

Select Badges

Select Supported Platforms

Select Status


dse (11) Versions 3.0.15

Installs/Configures Datastax Enterprise.

cookbook 'dse', '= 3.0.15', :supermarket
cookbook 'dse', '= 3.0.15'
knife supermarket install dse
knife supermarket download dse
Quality -%

Build Status

Datastax Enterprise Chef Cookbook (Apache Cassandra)

This cookbook installs and configures Datastax Enterprise. More info is here (DataStax Enterprise).

It uses officially released Datastax packages. It can tweak the Cassandra config files, but has no way of adding data or creating keyspaces in Cassandra (yet).


This cookbook is designed to be used in conjuction with a wrapper cookbook. Used alone, a single node cluster can be created, but in order to create a multiple node cluster a wrapper is recommended.

Example in a wrapper:

node.default['java']['jdk_version'] = "7"
node.default['cassandra']['seeds'] = ","
node.default['cassandra']['dse_version'] = "4.0.3-1"
node.default['cassandra']['max_heap_size'] = "12G"
node.default['cassandra']['heap_newsize'] = "1200M"

include_recipe "dse::cassandra"


This cookbook attempts to manage almost all Apache Cassandra configuration settings. It can also create Hadoop and Solr nodes, with less attribute to manage their config.

Apache Cassandra

This cookbook currently provides

  • Datastax 4.x.x (Datastax Enterprise Edition) via packages.


  • Chef 11 or higher

Supported OS Distributions

Tested on:

  • RHEL 6.3, 6.4
  • Ubuntu 14.04.1 LTS
  • Slight testing done on Ubuntu 12.04 (will require some edits)


The provided recipes are dse::cassandra, dse::solr, and dse::hadoop
* dse::cassandra will provision DSE as a cassandra node.
* dse::solr will provision DSE with solr enabled.
* dse::hadoop will provision DSE with hadoop enabled.

There are also recipes that should not be called directly that are used for configuration.
* dse::default sets up the templates
* dse::datastax sets up the datastax repos
* dse::datstax-agent configures the datastax-agent if needed
* dse::ssl (work in progress) sets up SSL keys on all nodes


This cookbook will install DSE Cassandra by default. Other attributes you can set are:


overall settings

  • node["cassandra"]["cluster_name"] (default: Test Cluster): The name of the cluster to provision
  • node["cassandra"]["vnodes"] (default: true): enable or disable vnodes
  • node["cassandra"]["intial_token"] (default: nil): the initial token to use. leave blank for vnodes
  • node["cassandra"]["num_tokens"] (default: 256): set the number of tokens to use
  • node["cassandra"]["solr"] (default: false): enable solr or not
  • node["cassandra"]["hadoop"] (default: false): enable hadoop or not

  • node["cassandra"]["dse_version"] (default: 4.0.3-1): dse version to install

  • node["cassandra"]["user"] (default: cassandra): the cassandra user

  • node["cassandra"]["group"] (default: cassandra): the cassandra group

cassandra.yaml settings

  • node["cassandra"]["listen_address"] (default: node['ipaddress']): the ipaddress to use for listen address
  • node["cassandra"]["rpc_address"] (default: node['ipaddress']): the ipaddress to use for rpc address
  • node["cassandra"]["broadcast_address"] (default: nil): the ipaddress to use for broadcast address
  • node["cassandra"]["seeds"] (default: node['ipaddress']): the ipaddress to use for the seed list
  • node["cassandra"]["concurrent_reads"] (default: 32): concurrent reads setting
  • node["cassandra"]["concurrent_writes"] (default: 32): concurrent writes setting
  • node["cassandra"]["compaction_thruput"] (default: 16): limit the throughput of compactions
  • node["cassandra"]["multithreaded_compaction"] (default: false): enable or disable multithreaded compaction
  • node["cassandra"]["in_memory_compaction_limit"] (default: 64): size limit for in-memory compactions
  • node["cassandra"]["trickle_fsync"] (default: false): enable trickle fsync, usually for ssd
  • node["cassandra"]["range_request_timeout_in_ms"] (default: 10000): default timeout on range requests
  • node["cassandra"]["thrift_framed_transport_size_in_mb"] (default: 15): the max size of a thrift frame
  • node["cassandra"]["thrift_max_message_length_in_mb"] (default: nil): the max message length of a thrift call
  • node["cassandra"]["concurrent_compactors"] (default: nil): the number of concurrent compactors to allow

Role based seed selection

  • node["cassandra"]["role_based_seeds"] (default: false): set to true to assign seeds based on members of dse-seed role
  • node['cassandra']['seed_role'] (default: role:dse-seed): set to a diffrent role to select seeds

gc settings

  • node["cassandra"]["CMSInitiatingOccupancyFraction"] (default: 65): cms occupancy fraction to use for gc
  • node["cassandra"]["max_heap_size"] (default: 8192M): default max heap size for cassandra
  • node["cassandra"]["heap_newsize"] (default: 800M): default new gen size for heap

authentication settings

  • node["cassandra"]["authentication"] (default: false): enable or disable authentication
  • node["cassandra"]["authorization"] (default: false): enable or disable authorization
  • node["cassandra"]["authenticator"] (default: ``): the authenticator to use (eg org.apache.cassandra.auth.AllowAllAuthenticator)
  • node["cassandra"]["authorizor"] (default: ``): the authorizor to use (eg org.apache.cassandra.auth.AllowAllAuthorizer)

audit logs

  • node["cassandra"]["log_level"] (default: INFO): the log level for cassandra (or solr/hadoop)
  • node["cassandra"]["audit_logging"] (default: false): turn on audit logging
  • node["cassandra"]["audit_dir"] (default: /var/log/cassandra): the directory to put audit logs in
  • node["cassandra"]["active_categories"] (default: ADMIN,AUTH,DDL,DCL): the categories to audit on


  • node["cassandra"]["dse"]["delegated_snitch"] (default: org.apache.cassandra.locator.SimpleSnitch): the snitch to use for dse
  • node["cassandra"]["dse"]["snitch"] (default: com.datastax.bdp.snitch.DseDelegateSnitch): the snitch to use in dse.yaml
  • node["cassandra"]["dse"]["service_name"] (default: dse): the name of the service
  • node["cassandra"]["dse"]["conf_dir"] (default: /etc/dse): the directory of dse config files
  • node["cassandra"]["dse"]["repo_user"] (default: ``): the datastax username for the repo
  • node["cassandra"]["dse"]["repo_pass"] (default: ``): the datastax password for the repo
  • node["cassandra"]["dse"]["rhel_repo_url"] (default: http://#{node['cassandra']['dse']['repo_user']}:#{node['cassandra']['dse']['repo_pass']} the rhel repo
  • node["cassandra"]["dse"]["debian_repo_url"] (default: http://#{node['cassandra']['dse']['repo_user']}:#{node['cassandra']['dse']['repo_pass']} the debian repo


  • node["hadoop"]["max_heap_size"] (default: 10G): the heap size for hadoop
  • node["hadoop"]["heap_newsize"] (default: 800M): the heap newgen size for hadoop
  • node["hadoop"]["map_child_java_opts"] (default: 4G): the size of the map child java heap
  • node["hadoop"]["reduce_child_java_opts"] (default: 4G): the size of the reduce child java heap
  • node["hadoop"]["map_red_localdir"] (default: /data/mapredlocal): the directory to use for map/reduce
  • node["hive"]["scratch_dir"] (default: /data/hive): the directory to use for hive
  • node["hadoop"]["map_reduce_parallel_copies"] (default: 20): the number of map reduce copies
  • node["hadoop"]["mapred_tasktracker_map_tasks_max"] (default: 23): the max number of map tasks
  • node["hadoop"]["mapred_tasktracker_reduce_tasks_max"] (default: 12): the max number of reduce tasks
  • node["hadoop"]["io_sort_mb"] (default: 512M): the size of iosort
  • node["hadoop"]["io_sort_factor"] (default: 64): the iosort factor


  • node["solr"]["max_heap_size"] (default: 14G): the heap size for solr
  • node["solr"]["heap_newsize"] (default: 2400M): the newgen heap size


These are generic java settings. Datastax recommends oracle java, so override openjdk default and download from a specific location.
* node["dse"]["manage_java"] (default: true): whether or not to use the java recipe to manage the java install
* node["java"]["install_flavor"] (default: oracle): the flavor of java to install
* node["java"]["jdk_version"] (default: 7): the version of java to use
* node['java']['jdk']['7']['x86_64']['url'] (default: ``): the url to get the java 7 file from


This portion is under construction. SSL does not currently 100% work.
* node["cassandra"]["dse"]["cassandra_ssl_dir"] (default: /etc/cassandra): the directory to use for pem files
* node["cassandra"]["dse"]["password_file"] (default: cassandra_pass.txt): the file to store the keystore pass in
* node["cassandra"]["dse"]["internode_encyption"] (default: none): the encyption to use (all, dc, rack)
* node["cassandra"]["dse"]["keystore"] (default: #{node["cassandra"]["dse"]["cassandra_ssl_dir"]}/#{node["hostname"]}.keystore): keystore name
* node["cassandra"]["dse"]["truststore"] (default: #{node["cassandra"]["dse"]["cassandra_ssl_dir"]}/#{node["hostname"]}.truststore): truststore name


These attributes are used to conigure the datastax-agent. This is used with Datastax Opscenter.

  • node["datastax-agent"]["enabled"] (default: false): whether to install the datastax agent and configure
  • node["datastax-agent"]["version"] (default: 4.1.1-1): the version of the datastax agent to install
  • node["datastax-agent"]["conf_dir"] (default: /var/lib/datastax-agent/conf): where the datastax-agent conf file is
  • node["datastax-agent"]["opscenter_ip"] (default: the Opscenter IP to connect to


  • java
  • yum
  • apt

Datastax recommends to use the Oracle jdk version. You can do this by setting an attribute in your environment or run list.

Currently, Oracle prevents you from downloading the package from their website, put it in Artifactory or something as a workaround.
You can override the java url with an attribute, show below.

Released under the Apache 2.0 License.

Dependent cookbooks

java ~> 1.14
yum ~> 3.5
yum-epel ~> 0.6
apt ~> 2.0

Contingent cookbooks

There are no cookbooks that are contingent upon this one.

CHANGELOG for dse cookbook

This file is used to list changes made in each version of the dse cookbook.


  • restart the datastax-agent on new version


  • added support to set MaxTenuringThreshold in cassandra-env


  • refactored some version checks
  • added role based seed assignment


  • add templates and upgrade to 4.5.2


  • add templates and upgrade to 4.0.4


  • remove the ssd tuning from this recipe


  • minor updates


  • adding ability to tune memtable thresholds


  • removed specific tuning to move it to os-tuning cookbook
  • adding ability to set concurrent compactors


  • added thrift frame size settings for hadoop requests


*adding recommended datastax tuning settings


  • adding cassandra specific gc settings


  • more hadoop tunung


  • adding hadoop tuning settings


  • First pass of node-to-node ssl
  • Adding another hadoop attribute
  • fixing hadoop map reduce dir, as hadoop didnt create
  • adding changes to ssl
  • adding a version-specific dse script
  • adding support to stop dse before an upgrade
  • hive scratch directory support


  • subscribed the dse service to java, so it will restart if java version changes
  • added more chefspec
  • moved the start of the dse service until after all the templates are set up


  • allow support for gossipingPropertyFileSnitch


  • Tell the OS that SSDs are present


  • Allow Solr and Hadoop Heap to be set dynamically


  • Allows this recipe to install the datastax-agent


  • Allow DSE 4.0 to be installed


  • Rename the cookbook to dse


  • Added kitchen tests
  • Added multiple data directory support


  • Initial release of cassandra

Check the Markdown Syntax Guide for help with Markdown.

The Github Flavored Markdown page describes the differences between markdown on github and standard markdown.

No quality metric results found