Vertica Blog

What’s New in Vertica 9.0?

This blog post was authored by Soniya Shah. In Vertica 9.0, we introduce new functionality including: • Eon Mode Beta • Supported Platform Updates • Machine Learning Enhancements • Apache Hadoop Integration Updates • Partition Grouping and Hierarchical Partitioning • Browsing S3 Data Using External Tables • Support for the UUID Data Type Eon Mode […]

CPU and Memory Starvation

This blog post was authored by Sumeet Keswani. What is Spread? Vertica uses an open source toolkit, Spread, to provide a high-performance control message service. Spread daemons start automatically when your database starts up for the first time. The spread daemons run on control nodes in your cluster. The control nodes manage message communication. On […]

Effective vsql in Vertica

This blog post was authored by Maurizio Felici. vsql is included in each Vertica installation and is lightweight, with a tight integration with Vertica. Vsql is installed on every Vertica server and can also be installed on non-server hosts using the client package. Executing SQL commands through vsql is often faster than navigating GUI’s menu. […]

Adding Nodes to Fault Groups

This blog post was authored by Sarah Lemaire. Suppose you are adding new cluster nodes to your Vertica database. You want to add those nodes to particular fault groups without having to restart your Vertica database. The following steps use the example of a database with five racks and fault groups, with 9 Vertica nodes […]

Analytic Queries in Vertica

This blog post was authored by Soniya Shah. Analytic functions handle complex analysis and reporting tasks. Here are some example use cases for Vertica analytic functions: • Rank the longest standing customers in a particular state • Calculate the moving average of retail volume over a specific time • Find the highest score among all […]

Integrating with Apache Spark

This blog post was authored by Soniya Shah. The Vertica Connector for Apache Spark is a fast parallel connector that allows you to use Apache Spark for pre-processing data. Apache Spark is an open-source, general purpose, cluster-computing framework. The Spark framework is based on Resilient Distributed Datasets (RDDs), which are logical collections of data partitioned […]

Working with Joins

This blog post was authored by Soniya Shah. Vertica supports a variety of join types. This post discusses the following joins: • Inner joins • Left, right, and full outer joins • Natural joins • Cross joins In Vertica, we refer to the tables participating in the join as left or right. The left table […]

Time Series Analytics

This blog post was authored by Soniya Shah. Time series analytics is a powerful Vertica tool that evaluates the values of a given set of variables over time and groups those values into a window based on a time interval for analysis and aggregation. Time series analytics is useful when you want to analyze discrete […]

Building a Secure Vertica Environment

This blog post was authored by Soniya Shah. Vertica has a client-server architecture system, where applications that reside on the client access the Vertica cluster through drivers including ODBC, JDBC, OLEDB and ADO.NET. This post discusses secure client to server communications, authenticating access to Vertica, and administrator access. Method Vertica Options Authentication: Validate user credentials […]

Vertica Presentation at the db tech showcase Tokyo 2017

On September 5th, Kanako Obayashi from the Vertica Best Practices team presented at the db tech showcase Tokyo 2017, one of the largest database events in Japan. Kanako’s presentation was about Vertica advanced analytics, including machine learning and geospatial analysis. More than 50 people attended her session. Kanako began her session by noting that more […]