[Hawkular-dev] more openshift issues

John Mazzitelli mazz at redhat.com
Wed Feb 22 08:00:27 EST 2017


Here's what I have so far:

https://github.com/jmazzitelli/hawkular-openshift-agent/blob/hack-os-start/hack/start-openshift.sh

which uses:

https://github.com/jmazzitelli/hawkular-openshift-agent/blob/hack-os-start/hack/env-openshift.sh

All my pods are blue. But I had to run it overnight - metrics took FOREVER for its ready probe to show "ready". But they are all blue now - says everything is running.

But I have no idea if things are working because I have no way to interact with my pods because: I have no valid routes working, I can't look at pod logs, and I can't use a pod terminal to probe around.

And people wonder why I don't upgrade things. Been days now, and I still can't get the latest OpenShift to work like it did before. :/

Here's the problems:

1. I can no longer get to the Hawkular Metrics URL from my browser or HawkFX:

$ curl https://hawkular-metrics.example.com/
curl: (6) Could not resolve host: hawkular-metrics.example.com

UI Console has a message about this: "The route is not accepting traffic yet because it has not been accepted by a router."

2. The same thing with my HOSA route - it doesn't work either with the same warning about not being accepted by a router. But I have a route defined - I can see it in my UI. Route named "hawkular-openshift-agent" has Hostname "http://hawkular-openshift-agent-default.router.default.svc.cluster.local/status " and Route To service "hawkular-openshift-agent". That service is defined, too.

It think the way routes are defined might have changed from 1.5-alpha2 to 1.5-alpha3.

3. I can't see any Logs for any pod. For example, I used to be able to see logs for HOSA but now all the Logs tab says, "The logs are no longer available or could not be loaded."

4. I can no longer use the Terminal tab in the UI Console. All pod terminals (including HOSA's) page in the UI say "Could not connect to the container. Do you have sufficient privileges?"

Note that my "admin" user has "cluster-admin" role via "oc adm policy add-cluster-role-to-user cluster-admin admin" so it should see everything (at least, that's how it worked in the older version).

----- Original Message -----
> On 02/21/2017 11:25 PM, John Mazzitelli wrote:
> > If I start openshift with "sudo ./openshift start" and then try to log in
> > like this:
> >
> >    oc login -u system:admin
> >
> > What would cause this:
> >
> >    Authentication required for https://192.168.1.15:8443 (openshift)
> >    Username: system:admin
> >    Password:
> >    error: username system:admin is invalid for basic auth
> >
> > When I start with "oc cluster up" I do not get asked for a password and it
> > "just works"
> 
> It seems I forgot to mention that. You need to use a "kubeconfig",
> created during the first boot. Most of the `oc` commands accept a config
> as parameter, but the easiest is to export a KUBECONFIG pointing to the
> admin file.
> 
> I have this on my ~/.bashrc :
> 
> export
> KUBECONFIG="/home/jpkroehling/go/src/github.com/openshift/origin/_output/local/bin/linux/amd64/openshift.local.config/master/admin.kubeconfig"
> 
> - Juca.
> _______________________________________________
> hawkular-dev mailing list
> hawkular-dev at lists.jboss.org
> https://lists.jboss.org/mailman/listinfo/hawkular-dev
> 


More information about the hawkular-dev mailing list