- Amazon
- Pivotal
- 2ndQuadrant
- Microsoft
- Timescale
- HighGo
- Data Dog
- Fivetran
- BlueTalon
- Goldman Sachs
- forhims
- Immuta
On occasion, professional developers will drop into the Postgresql.org mailing lists, meetups, and conferences to ask the question, “Why isn’t PostgreSQL development on Github?” In an effort to see if the demand was really there and not just anecdotal we ran a poll/survey over several social media platforms that asked a simple question:
Should PostgreSQL development move to Github?
We received well over 300 responses and the majority (75%+) chose a move to Github or to something like Github. This was an unscientific poll but it does point out a few interesting topics for consideration:
This poll does not answer point #3; it only provides a data point that people may desire a modern collaboration platform. The key takeaway from the conversation about migrating to Github or similar service is the future generation of developers use technology such as Slack and Microsoft Teams. They expect a bug/issue tracker. They demand simplicity in collaboration and most importantly they will run a cost->benefit analysis to determine if the effort to contribute is a net positive.
It should also be considered that this is not just individual potential contributors. There are many corporations big and small that rely on the success of PostgreSQL. Those corporations will not contribute as much directly to PostgreSQL if the cost to benefit analysis is a net negative. They will instead contribute through other more productive means that produce a net positive when the cost->benefit analysis is run. A good example of this analysis is the proliferation of external projects such as pg_auto_failover, patroni and lack of direct contribution from innovative extension based companies.
There are those within the Postgresql.org community that would suggest that we do not need a culture shift within PostgreSQL but that does not take into account the very clear market dynamics that are driving the growth of PostgreSQL, Postgres, and the global ecosystem. It is true that 20 years of hard work by Postgresql.org started the growth and it is also true that the majority of growth in the ecosystem and community is from products such as Greenplum, Aurora, Azure, and Timescale. The growth in the ecosystem is from the professional community and that ecosystem will always perform a cost to benefit analysis before contributing.
It is not that we should create radical rifts or disrupt our culture. It is to say that we must evolve and shift our community thinking. We need to be able to consider the big picture. A discussion should never start as an opposition to change. The idea of change should be an open discussion about possibility and vision. It should always include whether the change is a good idea and it should always avoid visceral reactions of, "works for me,” “no,” or “we tried that 15 years ago." Those reactions are immature and lacking in the very thing the community needs to continue to grow: positivity, inclusion, vision, and inspiration.
Pivotal Sponsor Highlight Blog for PostgresConf 2019
Written by:
Jacque Istok, Head of Data, Pivotal
1. Greenplum has its own community; what do you hope to achieve by joining the Postgres community and PostgresConf?
Both interest and adoption of Postgres have skyrocketed over the last two years, and we feel fortunate to be a part of the extended community. We have worked very hard to uplevel the base version of Postgres within Greenplum to more current levels and to be active in the Postgres community. We see Greenplum as a parallel (and analytic focused) implementation of Postgres, and we encourage the community to continue to embrace both the technology and the goal of the Greenplum project, which is Postgres at scale.
2. Are you planning to provide any new tech (PG features, etc.)?
This year we plan to announce several new things for both Greenplum and Postgres. We’re introducing new innovations in our cloud offerings in the marketplaces of AWS, Azure, and GCP. We also have major news about both our natural language at-scale analytics solution based on Apache Solr, and our multi-purpose machine learning and graph analytics library Apache MADlib. The next major release of Greenplum is a major focus as well, differentiating Greenplum from each of its competitors and bringing us ever closer to the latest versions of Postgres.
3. Are there any rising stars in the community you’d like to give props to?
While it seems a little self-serving, I would like to take the opportunity to give props to the Pivotal Data Team. This team is a 300+ worldwide organization that helps our customers, our prospects, and the community to solve real world and really hard data problems—solved in part through Postgres technology. They all attack these use cases with passion and truly make a difference in the lives of the people that their solutions touch. I couldn’t wish to work with a finer group.
4. What is the number one benefit you see within Postgres that everyone should be aware of?
The number one benefit of Postgres is really its flexibility. This database chameleon can be used for SQL, NoSQL, Big Data, Microservices, time series data, and much more. In fact, our latest analytic solution, MADLib Flow, leverages Postgres as an operational engine. For example, Machine Learning models created in Greenplum can be pushed into a restful API as part of an agile continuous integration/continuous delivery pipeline easily and efficiently—making Postgres the power behind what I still like to think of as #DataOps.
5. What is the best thing about working with the Postgres community?
I deeply admire the passion and consistency of the community behind Postgres, constantly and incrementally improving this product over decades. And because Greenplum is based on Postgres, we get to interact with this vast community of talent. We are also able to more seamlessly interact with ecosystem products that already work with Postgres, making the adoption of Greenplum that much easier.
6. Tell us why you believe people should attend PostgresConf 2019 in March.
PostgresConf is going to be awesome, and I can’t wait for it to start! With Pivotal, Amazon, and EnterpriseDB headlining as Diamond sponsors, Greenplum Summit (along with multiple other summits), and high-quality speakers and content across the board, this year’s PostgresConf promises to be bigger and better than ever and surely won’t disappoint.
We’re thrilled to be back to present the second annual Greenplum Summit on March 19th at PostgresConf. Our theme this year is “Scale Matters”, and what we’ve seen with our customers is that every year it matters more and more. Our users are part of organizations that are generating tons of data and their need to easily and quickly ingest and interrogate all of it is paramount. This is true even more now than ever before as the insights that can be found not only help differentiate them from their competitors, but are also used to build better products and increase customer loyalty.
The day will be filled with real-world case studies from Greenplum users including Morgan Stanley, the European Space Astronomy Centre, Insurance Australia Group, Purdue University, Baker Hughes (a GE company), Conversant, and others, plus presentations and deep-dive tech sessions for novices and experts alike.
As one of the Founders of United States PostgreSQL, a former Director of Software in the Public Interest (one of the NPOs behind Postgresql.org), a former committer (web), former major contributor, President of the oldest PostgreSQL company still independent in North America, and the Founder of Postgres Conference (in the U.S.), I thought I would offer a knowledgeable perspective.
I have had long discussions with one of the primary people within the Fundacion PostgreSQL (Alvaro) and his heart is in the best interest of the community, even if Postgresql.org, PGEU and PGCAC do not agree. You can see this demonstrated within Fundacion’s trademark policy. That said, Fundacion PostgreSQL did go about their actions in an incorrect way. There should have been an open discussion and they should have provided PGCAC the opportunity to resolve the trademark issues on their own. It is also true that while I believe PGEU and PGCAC believe they are protecting the community, if they were interested in positive community growth and collaboration, they would not be taking the approach they currently are. The current path has far reaching implications that PGEU and PGCAC do not see.
Further, the PostgreSQL Community Association of Canada and Fundacion PostgreSQL have resorted to terrible language in representing what is actually going on within the disagreement. Using language such as, “An attack on our community” or “PostgreSQL attacks the community” is immature at best and at worst an intentional decision to use good faith and mindshare against what is largely just a disagreement that could be solved with an active mediator and a few phone calls. If this disagreement is about the best interest of the PostgreSQL community, shouldn’t that involve discourse, honesty, transparency, and kind communication?
(1) Internet consulting.
(2) Internet presence provider- DNS hosting.
(3) Commercial internet support for database applications development and implementation including the ability to host internet domains (as an internet service provider) and provide a wide range of web site development, programming and information technology services, namely computer software architecture, design and/or development services.
(4) Computer hardware sales and service.
The solution to the whole problem is simple; a single contract that states:
The contract should not contain language in regards to future potential filings that involve but are not exclusive to the word Postgres or PostgreSQL. There are already a number of filings worldwide that use Postgres or PostgreSQL as part of an overall mark inclusively such as Postgres Pro, Postgres Plus, Postgres Always On and Postgres Enterprise Manager, all of which are not owned but PGCAC or PGEU.
Because it is the right thing to do. Otherwise this whole affair is going to end up costing one entity or another way too much money for no purpose. There is no clear distinction on who would legally win, and in either situation the main sufferers are the PostgreSQL community. Let’s have the parties show an act of kindness for the betterment of everyone involved.
We are having yet another PGConf Mini in NYC. The event is scheduled for December 14th, 2017 and Work Bench is hosting:
Efficiently and Safely Propagate Data Changes Without Triggers!
After National 2017, the PGConf US team decided to take a long hard road to bring the best in educational content and advocacy to the entire Postgres ecosystem. Today, we welcome 2ndQuadrant as a Platinum sponsor and thank them for helping us with our year over year projects.


With the pending GA release of Star Link, more cities adopting municipal WIFI, and the growth of remote work due to the pandemic, one would assume that cell phone providers would be falling hand over fist to provide quality services at a quality price. Unfortunately this is not the case and it is hurting the future of our workforce.

Carriers currently think that 30GB of Hotspot data is enough for the average digital nomad. While that may have been the case previously, it isn’t any longer. There is a hundreds-of-thousands-strong workforce specializing in Information Technology, Digital Design, Web Development, and other creative industries. They are living, working and adventuring in RVs, Sprinters, Skooolies (pictured), and other vehicle dwelling options. This community is nomadic by design and should not be artificially limited by technology.
Why is it that we are being charged so much money for so little? Outside of two providers announcing rural initiatives, we are still acting like data needs are small and that they come at a quality price. Data is ubiquitous; it is the new water (for getting work done). In the new economy workers need faster, higher quality, and unlimited amounts of data to get their job done.
Where could you go and what could you experience if you weren’t tethered by the faux limitations of bandwidth?
Henrietta Dombrovskaya, Diversity, Equity, and Inclusion member for PostgresConf as well as Chicago Postgresql organizer nominated for Technologist of the Year!
A large, well known media company is seeking a Senior Level PostgreSQL Engineer and Architect. This is an on-site engagement, however the company is known to be lifestyle friendly with reasonable working hours, good pay, and benefits. Specific talents requested are the ability to mentor. The location is Seattle, WA. If you are interested in this position please contact randy@neuringerco.com with your resume.
We wanted to highlight some of the fantastic work that is being done by the ecosystem with Postgres Extensions. Although the base of Postgres is the amazing and extensible PostgreSQL, a lot of users don’t realize that Postgres has the feature they are looking for, if only they were to look to the ecosystem.
pgaudit : The goal of pgAudit is to provide PostgreSQL users with the capability to produce audit logs often required to comply with government, financial, or ISO certifications.
pg_credereum : pg_credereum is a PostgreSQL extension that provides a cryptographically verifiable audit capability for a PostgreSQL database, bringing some properties of blockchain to relational DBMS.
H3-pg : PostgreSQL bindings for H3, a hierarchical hexagonal geospatial indexing system.
There are a ton of Postgres compatible features out there. Some of them are overlooked core features and some of them require installing a different version of Postgres. Here are a few examples:
TimescaleDB: Time series data management with Postgres
YugabyteDB: Globally Distributed database with PostgreSQL compatibility
Postgres-XL: Horizontally partitioned PostgreSQL
Agensgraph: Graph capabilities with Postgres
PG-Strom: GPU accelerated extension for Postgres