So I built GPDB 5.0.alpha (including ORCA) and brought up the database on my Dev machine. Great start.I then tar-gzipped up /usr/local/gpdb and copied it to another CentOS 6.8 server for deployment. This went well until I attempted to initialize Greenplum, then I ran into errors:20170405:07:26:47:009622 gpinitsystem:mdw:gpadmin-[INFO]:-Starting the Master in admin modeError: unable to import module: No module named psutil20170405:07:26:50:gpinitsystem:mdw:gpadmin-[FATAL]:-Unknown host Script Exiting!20170405:07:26:50:009622 gpinitsystem:mdw:gpadmin-[WARN]:-Script has left Greenplum Database in an incomplete stateSo I attempted to install the psutil python module, but that failed because this is not a Dev machine with all of the dependencies installed:psutil/_psutil_linux.c:12:20: error: Python.h: No such file or directorypsutil/_psutil_linux.c:40: error: ‘CHAR_BIT’ undeclared here (not in a function)How are companies deploying the OSS version of GPDB on servers that are not full Dev boxes? Or does setting up their "production" clusters involve building them as essentially Dev servers with the full list of -devel packages installed in order to run OSS GPDB?
--
You received this message because you are subscribed to the Google Groups "Greenplum Developers" group.
To unsubscribe from this group and stop receiving emails from it, send an email to gpdb-dev+unsubscribe@greenplum.org.
Hey Keaton,
Try installing Python development libraries, yum install python-devel. This should help.
Personally I feel that we need to coalesce around just a few
deployment/packaging
options. Our Python scripts are not going anywhere -- so that's option #1 that's
here to stay. As for option #2 I felt that what EPAM did was actually
the right thing
to do. GPDB existing independently of the rest of the IT
infrastructure is getting
more rare these days. All those accounts are running things like
Puppet/Chef/Ansible
and demand that THAT is the single source of truth for configuration
and deployment
of every single node in that IT department (not some random
configuration files of
individual software packages).
I would really like to encourage GPDB community to consider this as
option #2 and
help maintain Bigtop's puppet code.
Thoughts?
Thanks,
Roman.
Andreas Scherbaum
Principal Software Engineer
GoPivotal Deutschland GmbH
Amtsgericht Königstein im Taunus, HRB 8433
Geschäftsführer: Andrew Michael Cohen, Paul Thomas DacierOn Thu, Apr 6, 2017 at 12:30 PM, C.J. Jameson <cjam...@pivotal.io> wrote:
> My sense, given the codebase, is that it'd be very hard to identify a
> comprehensive API that's stable enough for you to rely on. Even if we
> characterized it 98%, the small missing parts would be unknown and risks.
> Greenplum's "surface area", so to speak, isn't well organized like an API, I
> feel, because of the various entry points.
I think we mean different things when we say "API". What I mean by this
is literally, GPDB taking responsibility of maintaining a few scripts that
Bigtop now has to maintain on its own (and risk breakage with every
release of GPDB). These are far from rocket science:
https://github.com/apache/bigtop/tree/master/bigtop-deploy/puppet/modules/gpdb/templates
Basically all of the files under there aside from gpssh.conf and gp_dbid
Take a look for yourself - those are trivial. In fact, we can even merge them
into a single uber script on the GPDB side with actions that a system
like Puppet expects.
That script IS the API I'm suggesting here -- nothing more.
> Rather than prescribing the API and trying to stabilize it, we can describe
> the needs of BigTop and automate checks that it still works. So the idea is:
> codify BigTop smoke-testing as a downstream check in our pipeline ... not
> necessarily to block further pushes but to know when it breaks.
Sure. Testing comes next and Bigtop would be more than happy to participate,
but lets start with the basics -- GPDB offering a single script with a fixed set
of actions as a stable API for all consumers like Puppet/Chef/Ansible/etc.