[JBoss JIRA] (WFCORE-1521) Autostart order in domain mode
by Guillermo González de Agüero (JIRA)
[ https://issues.jboss.org/browse/WFCORE-1521?page=com.atlassian.jira.plugi... ]
Guillermo González de Agüero updated WFCORE-1521:
-------------------------------------------------
Description:
When defining servers in domain mode, they can be marked to be autostarted, but an order of starting cannot be specified (they'll start in parallel). A new option could be added to start one server only after others have already started. A server start would be posponed until all servers with lower order have already finished.
An example use case would be a server that requires some resources of another one (JMS queues for example).
This can be easily achieved with a CLI batch command, but it would be simpler to have it available out of the box.
was:
When defining servers in domain mode, they can be marked to be autostarted, but an order of starting cannot be specified.
An example use case would be a server that requires some resources of another one (JMS queues for example).
This can be easily achieved with a CLI batch command, but it would be simpler to have it available out of the box.
> Autostart order in domain mode
> ------------------------------
>
> Key: WFCORE-1521
> URL: https://issues.jboss.org/browse/WFCORE-1521
> Project: WildFly Core
> Issue Type: Feature Request
> Reporter: Guillermo González de Agüero
>
> When defining servers in domain mode, they can be marked to be autostarted, but an order of starting cannot be specified (they'll start in parallel). A new option could be added to start one server only after others have already started. A server start would be posponed until all servers with lower order have already finished.
> An example use case would be a server that requires some resources of another one (JMS queues for example).
> This can be easily achieved with a CLI batch command, but it would be simpler to have it available out of the box.
--
This message was sent by Atlassian JIRA
(v6.4.11#64026)
8 years, 7 months
[JBoss JIRA] (WFCORE-1521) Autostart order in domain mode
by Guillermo González de Agüero (JIRA)
Guillermo González de Agüero created WFCORE-1521:
----------------------------------------------------
Summary: Autostart order in domain mode
Key: WFCORE-1521
URL: https://issues.jboss.org/browse/WFCORE-1521
Project: WildFly Core
Issue Type: Feature Request
Reporter: Guillermo González de Agüero
When defining servers in domain mode, they can be marked to be autostarted, but an order of starting cannot be specified.
An example use case would be a server that requires some resources of another one (JMS queues for example).
This can be easily achieved with a CLI batch command, but it would be simpler to have it available out of the box.
--
This message was sent by Atlassian JIRA
(v6.4.11#64026)
8 years, 7 months
[JBoss JIRA] (WFCORE-1503) Reading the logging subsystem resource can be slow if several log files exist
by James Perkins (JIRA)
[ https://issues.jboss.org/browse/WFCORE-1503?page=com.atlassian.jira.plugi... ]
James Perkins commented on WFCORE-1503:
---------------------------------------
[~pvorb] I believe you're the one that was having issues with this. If so and you're willing to test a patch grab the attached {{wildfly-logging-patch-02.jar}} JAR and place it in {{$JBOSS_HOME/modules/system/layers/base/org/jboss/as/logging/main/}} directory and change the {{module.xml}} in that file to point to the new library. From my local tests it seemed faster.
> Reading the logging subsystem resource can be slow if several log files exist
> -----------------------------------------------------------------------------
>
> Key: WFCORE-1503
> URL: https://issues.jboss.org/browse/WFCORE-1503
> Project: WildFly Core
> Issue Type: Bug
> Components: Logging
> Reporter: James Perkins
> Assignee: James Perkins
> Labels: 2.2
> Attachments: wildfly-logging-patch-02.jar
>
>
> The logging subsystem uses a dynamic resource to list log files as resources. These are the files that are allowed to be read or downloaded and are attached to known file handlers.
> When several file handlers are defined the file tree walker is invoked several times per {{read-resource-operation}}. This may be happening for each stage of an operation. Ideally we could either cache the file listing per-operation or only execute at the runtime stage. This may not be an option for dynamic resources however.
> As a result of this poor performance the web console is very slow to load the logging subsystem configuration. This is not an issue with the web console.
--
This message was sent by Atlassian JIRA
(v6.4.11#64026)
8 years, 7 months
[JBoss JIRA] (WFCORE-1503) Reading the logging subsystem resource can be slow if several log files exist
by James Perkins (JIRA)
[ https://issues.jboss.org/browse/WFCORE-1503?page=com.atlassian.jira.plugi... ]
James Perkins updated WFCORE-1503:
----------------------------------
Attachment: wildfly-logging-patch-02.jar
> Reading the logging subsystem resource can be slow if several log files exist
> -----------------------------------------------------------------------------
>
> Key: WFCORE-1503
> URL: https://issues.jboss.org/browse/WFCORE-1503
> Project: WildFly Core
> Issue Type: Bug
> Components: Logging
> Reporter: James Perkins
> Assignee: James Perkins
> Labels: 2.2
> Attachments: wildfly-logging-patch-02.jar
>
>
> The logging subsystem uses a dynamic resource to list log files as resources. These are the files that are allowed to be read or downloaded and are attached to known file handlers.
> When several file handlers are defined the file tree walker is invoked several times per {{read-resource-operation}}. This may be happening for each stage of an operation. Ideally we could either cache the file listing per-operation or only execute at the runtime stage. This may not be an option for dynamic resources however.
> As a result of this poor performance the web console is very slow to load the logging subsystem configuration. This is not an issue with the web console.
--
This message was sent by Atlassian JIRA
(v6.4.11#64026)
8 years, 7 months
[JBoss JIRA] (JBASMP-77) ConcurrentModificationException in deploy-artifact
by James Perkins (JIRA)
[ https://issues.jboss.org/browse/JBASMP-77?page=com.atlassian.jira.plugin.... ]
James Perkins commented on JBASMP-77:
-------------------------------------
Do you have a way to duplicate it? I can't seem to get it to fail locally.
> ConcurrentModificationException in deploy-artifact
> --------------------------------------------------
>
> Key: JBASMP-77
> URL: https://issues.jboss.org/browse/JBASMP-77
> Project: JBoss AS Maven Plugins
> Issue Type: Bug
> Components: deploy
> Affects Versions: 7.7.Final
> Reporter: Daniele Gaffuri
>
> In the same situation of JBASMP-57 deploy-artifact throws the following exception:
> {code}
> [ERROR] Failed to execute goal org.jboss.as.plugins:jboss-as-maven-plugin:7.7.SCE:deploy-artifact (deploy) on project sideup-integration: Could not execute goal deploy-artifact on null. Reason: null: ConcurrentModificationException -> [Help 1]
> org.apache.maven.lifecycle.LifecycleExecutionException: Failed to execute goal org.jboss.as.plugins:jboss-as-maven-plugin:7.7.SCE:deploy-artifact (deploy) on project sideup-integration: Could not execute goal deploy-artifact on null. Reason: null
> at org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:217)
> at org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:153)
> at org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:145)
> at org.apache.maven.lifecycle.internal.LifecycleModuleBuilder.buildProject(LifecycleModuleBuilder.java:84)
> at org.apache.maven.lifecycle.internal.LifecycleModuleBuilder.buildProject(LifecycleModuleBuilder.java:59)
> at org.apache.maven.lifecycle.internal.LifecycleStarter.singleThreadedBuild(LifecycleStarter.java:183)
> at org.apache.maven.lifecycle.internal.LifecycleStarter.execute(LifecycleStarter.java:161)
> at org.apache.maven.DefaultMaven.doExecute(DefaultMaven.java:320)
> at org.apache.maven.DefaultMaven.execute(DefaultMaven.java:156)
> at org.apache.maven.cli.MavenCli.execute(MavenCli.java:537)
> at org.apache.maven.cli.MavenCli.doMain(MavenCli.java:196)
> at org.apache.maven.cli.MavenCli.main(MavenCli.java:141)
> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
> at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> at java.lang.reflect.Method.invoke(Method.java:606)
> at org.codehaus.plexus.classworlds.launcher.Launcher.launchEnhanced(Launcher.java:290)
> at org.codehaus.plexus.classworlds.launcher.Launcher.launch(Launcher.java:230)
> at org.codehaus.plexus.classworlds.launcher.Launcher.mainWithExitCode(Launcher.java:409)
> at org.codehaus.plexus.classworlds.launcher.Launcher.main(Launcher.java:352)
> Caused by: org.apache.maven.plugin.MojoExecutionException: Could not execute goal deploy-artifact on null. Reason: null
> at org.jboss.as.plugin.deployment.AbstractDeployment.doExecute(AbstractDeployment.java:159)
> at org.jboss.as.plugin.deployment.AbstractDeployment.execute(AbstractDeployment.java:112)
> at org.apache.maven.plugin.DefaultBuildPluginManager.executeMojo(DefaultBuildPluginManager.java:101)
> at org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:209)
> ... 19 more
> Caused by: java.util.ConcurrentModificationException
> at java.util.LinkedHashMap$LinkedHashIterator.nextEntry(LinkedHashMap.java:394)
> at java.util.LinkedHashMap$KeyIterator.next(LinkedHashMap.java:405)
> at java.util.AbstractCollection.addAll(AbstractCollection.java:341)
> at org.jboss.as.plugin.deployment.DeployArtifact.validate(DeployArtifact.java:84)
> at org.jboss.as.plugin.deployment.AbstractDeployment.doExecute(AbstractDeployment.java:136)
> ... 22 more
> {code}
> This is probably due to changes in related commit [37bc2c1], causing both deploy and undeploy validate methods to add the set of dependencies to itself:
> {code}
> final Set<Artifact> dependencies = project.getDependencyArtifacts();
> // Allows provided dependencies to be seen
> dependencies.addAll(project.getDependencyArtifacts());
> {code}
> {code}
> final Set<Artifact> dependencies = project.getDependencyArtifacts();
> dependencies.addAll(this.project.getDependencyArtifacts());
> {code}
--
This message was sent by Atlassian JIRA
(v6.4.11#64026)
8 years, 7 months
[JBoss JIRA] (DROOLS-1143) Guice error resolving implementation for RepositorySystem in shaded jar for kie-ci 6.3.0.Final with spark streaming (1.6.1)
by Lucie Zhao (JIRA)
[ https://issues.jboss.org/browse/DROOLS-1143?page=com.atlassian.jira.plugi... ]
Lucie Zhao commented on DROOLS-1143:
------------------------------------
I got this working, it was a class loader issue, please close this when you have time.
> Guice error resolving implementation for RepositorySystem in shaded jar for kie-ci 6.3.0.Final with spark streaming (1.6.1)
> ---------------------------------------------------------------------------------------------------------------------------
>
> Key: DROOLS-1143
> URL: https://issues.jboss.org/browse/DROOLS-1143
> Project: Drools
> Issue Type: Bug
> Reporter: Lucie Zhao
> Assignee: Mario Fusco
>
> I'm working on a project where I'm trying to integrate drools with spark streaming. I've seen this error on DROOLS-723 where it's reported to have been fixed in 6.2.0.Final but I'm using 6.3.0.Final and it's not working for me. I am able to run a quick test in my eclipse work space without the spark piece and it is working for me.
> here's the stack trace:
> {code}
> 016-04-27 10:22:32,360 WARN [streaming-job-executor-0] Sisu (Logs.java:warn(394)) - Error injecting: org.apache.maven.execution.DefaultMavenExecutionRequestPopulator
> com.google.inject.ProvisionException: Guice provision errors:
> 1) No implementation for org.apache.maven.repository.RepositorySystem was bound.
> while locating org.apache.maven.execution.DefaultMavenExecutionRequestPopulator
> 1 error
> at com.google.inject.internal.InjectorImpl$3.get(InjectorImpl.java:974)
> at com.google.inject.internal.InjectorImpl.getInstance(InjectorImpl.java:1000)
> at org.eclipse.sisu.space.AbstractDeferredClass.get(AbstractDeferredClass.java:48)
> at com.google.inject.internal.ProviderInternalFactory.provision(ProviderInternalFactory.java:84)
> at com.google.inject.internal.InternalFactoryToInitializableAdapter.provision(InternalFactoryToInitializableAdapter.java:52)
> at com.google.inject.internal.ProviderInternalFactory$1.call(ProviderInternalFactory.java:70)
> at com.google.inject.internal.ProvisionListenerStackCallback$Provision.provision(ProvisionListenerStackCallback.java:100)
> at org.eclipse.sisu.plexus.PlexusLifecycleManager.onProvision(PlexusLifecycleManager.java:133)
> at com.google.inject.internal.ProvisionListenerStackCallback$Provision.provision(ProvisionListenerStackCallback.java:108)
> at com.google.inject.internal.ProvisionListenerStackCallback.provision(ProvisionListenerStackCallback.java:55)
> at com.google.inject.internal.ProviderInternalFactory.circularGet(ProviderInternalFactory.java:68)
> at com.google.inject.internal.InternalFactoryToInitializableAdapter.get(InternalFactoryToInitializableAdapter.java:45)
> at com.google.inject.internal.ProviderToInternalFactoryAdapter$1.call(ProviderToInternalFactoryAdapter.java:46)
> at com.google.inject.internal.InjectorImpl.callInContext(InjectorImpl.java:1018)
> at com.google.inject.internal.ProviderToInternalFactoryAdapter.get(ProviderToInternalFactoryAdapter.java:40)
> at com.google.inject.Scopes$1$1.get(Scopes.java:59)
> at com.google.inject.internal.InternalFactoryToProviderAdapter.get(InternalFactoryToProviderAdapter.java:41)
> at com.google.inject.internal.InjectorImpl$3$1.call(InjectorImpl.java:965)
> at com.google.inject.internal.InjectorImpl.callInContext(InjectorImpl.java:1011)
> at com.google.inject.internal.InjectorImpl$3.get(InjectorImpl.java:961)
> at org.eclipse.sisu.inject.LazyBeanEntry.getValue(LazyBeanEntry.java:82)
> at org.eclipse.sisu.plexus.LazyPlexusBean.getValue(LazyPlexusBean.java:51)
> at org.codehaus.plexus.DefaultPlexusContainer.lookup(DefaultPlexusContainer.java:260)
> at org.codehaus.plexus.DefaultPlexusContainer.lookup(DefaultPlexusContainer.java:252)
> at org.codehaus.plexus.DefaultPlexusContainer.lookup(DefaultPlexusContainer.java:246)
> at org.kie.scanner.embedder.PlexusComponentProvider.lookup(PlexusComponentProvider.java:42)
> at org.kie.scanner.embedder.MavenEmbedder.buildMavenExecutionRequest(MavenEmbedder.java:111)
> at org.kie.scanner.embedder.MavenEmbedder.<init>(MavenEmbedder.java:84)
> at org.kie.scanner.embedder.MavenEmbedder.<init>(MavenEmbedder.java:75)
> at org.kie.scanner.embedder.MavenEmbedder.<init>(MavenEmbedder.java:69)
> at org.kie.scanner.embedder.MavenProjectLoader.parseMavenPom(MavenProjectLoader.java:55)
> at org.kie.scanner.embedder.MavenProjectLoader.parseMavenPom(MavenProjectLoader.java:49)
> at org.kie.scanner.ArtifactResolver.getResolverFor(ArtifactResolver.java:127)
> at org.kie.scanner.ArtifactResolver.getResolverFor(ArtifactResolver.java:90)
> at org.kie.scanner.KieRepositoryScannerImpl.setKieContainer(KieRepositoryScannerImpl.java:88)
> at org.drools.compiler.kie.builder.impl.KieServicesImpl.newKieScanner(KieServicesImpl.java:139)
> at com.emc.ctooil.spark.SparkKafkaDroolsConsumer$4.call(SparkKafkaDroolsConsumer.java:239)
> at com.emc.ctooil.spark.SparkKafkaDroolsConsumer$4.call(SparkKafkaDroolsConsumer.java:1)
> at org.apache.spark.streaming.api.java.JavaDStreamLike$$anonfun$foreachRDD$3.apply(JavaDStreamLike.scala:335)
> at org.apache.spark.streaming.api.java.JavaDStreamLike$$anonfun$foreachRDD$3.apply(JavaDStreamLike.scala:335)
> at org.apache.spark.streaming.dstream.DStream$$anonfun$foreachRDD$1$$anonfun$apply$mcV$sp$3.apply(DStream.scala:661)
> at org.apache.spark.streaming.dstream.DStream$$anonfun$foreachRDD$1$$anonfun$apply$mcV$sp$3.apply(DStream.scala:661)
> at org.apache.spark.streaming.dstream.ForEachDStream$$anonfun$1$$anonfun$apply$mcV$sp$1.apply$mcV$sp(ForEachDStream.scala:50)
> at org.apache.spark.streaming.dstream.ForEachDStream$$anonfun$1$$anonfun$apply$mcV$sp$1.apply(ForEachDStream.scala:50)
> at org.apache.spark.streaming.dstream.ForEachDStream$$anonfun$1$$anonfun$apply$mcV$sp$1.apply(ForEachDStream.scala:50)
> at org.apache.spark.streaming.dstream.DStream.createRDDWithLocalProperties(DStream.scala:426)
> at org.apache.spark.streaming.dstream.ForEachDStream$$anonfun$1.apply$mcV$sp(ForEachDStream.scala:49)
> at org.apache.spark.streaming.dstream.ForEachDStream$$anonfun$1.apply(ForEachDStream.scala:49)
> at org.apache.spark.streaming.dstream.ForEachDStream$$anonfun$1.apply(ForEachDStream.scala:49)
> at scala.util.Try$.apply(Try.scala:161)
> at org.apache.spark.streaming.scheduler.Job.run(Job.scala:39)
> at org.apache.spark.streaming.scheduler.JobScheduler$JobHandler$$anonfun$run$1.apply$mcV$sp(JobScheduler.scala:224)
> at org.apache.spark.streaming.scheduler.JobScheduler$JobHandler$$anonfun$run$1.apply(JobScheduler.scala:224)
> at org.apache.spark.streaming.scheduler.JobScheduler$JobHandler$$anonfun$run$1.apply(JobScheduler.scala:224)
> at scala.util.DynamicVariable.withValue(DynamicVariable.scala:57)
> at org.apache.spark.streaming.scheduler.JobScheduler$JobHandler.run(JobScheduler.scala:223)
> at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
> at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
> at java.lang.Thread.run(Thread.java:745)
> java.lang.RuntimeException: org.kie.scanner.embedder.MavenEmbedderException: com.google.inject.ProvisionException: Guice provision errors:
> 1) No implementation for org.apache.maven.repository.RepositorySystem was bound.
> while locating org.apache.maven.execution.DefaultMavenExecutionRequestPopulator
> at ClassRealm[maven, parent: ClassRealm[maven-parent, parent: null]]
> at ClassRealm[maven, parent: ClassRealm[maven-parent, parent: null]]
> while locating org.apache.maven.execution.MavenExecutionRequestPopulator
> 1 error
> role: org.apache.maven.execution.MavenExecutionRequestPopulator
> roleHint:
> at org.kie.scanner.embedder.MavenProjectLoader.parseMavenPom(MavenProjectLoader.java:57)
> at org.kie.scanner.embedder.MavenProjectLoader.parseMavenPom(MavenProjectLoader.java:49)
> at org.kie.scanner.ArtifactResolver.getResolverFor(ArtifactResolver.java:127)
> at org.kie.scanner.ArtifactResolver.getResolverFor(ArtifactResolver.java:90)
> at org.kie.scanner.KieRepositoryScannerImpl.setKieContainer(KieRepositoryScannerImpl.java:88)
> at org.drools.compiler.kie.builder.impl.KieServicesImpl.newKieScanner(KieServicesImpl.java:139)
> at com.emc.ctooil.spark.SparkKafkaDroolsConsumer$4.call(SparkKafkaDroolsConsumer.java:239)
> at com.emc.ctooil.spark.SparkKafkaDroolsConsumer$4.call(SparkKafkaDroolsConsumer.java:1)
> at org.apache.spark.streaming.api.java.JavaDStreamLike$$anonfun$foreachRDD$3.apply(JavaDStreamLike.scala:335)
> at org.apache.spark.streaming.api.java.JavaDStreamLike$$anonfun$foreachRDD$3.apply(JavaDStreamLike.scala:335)
> at org.apache.spark.streaming.dstream.DStream$$anonfun$foreachRDD$1$$anonfun$apply$mcV$sp$3.apply(DStream.scala:661)
> at org.apache.spark.streaming.dstream.DStream$$anonfun$foreachRDD$1$$anonfun$apply$mcV$sp$3.apply(DStream.scala:661)
> at org.apache.spark.streaming.dstream.ForEachDStream$$anonfun$1$$anonfun$apply$mcV$sp$1.apply$mcV$sp(ForEachDStream.scala:50)
> at org.apache.spark.streaming.dstream.ForEachDStream$$anonfun$1$$anonfun$apply$mcV$sp$1.apply(ForEachDStream.scala:50)
> at org.apache.spark.streaming.dstream.ForEachDStream$$anonfun$1$$anonfun$apply$mcV$sp$1.apply(ForEachDStream.scala:50)
> at org.apache.spark.streaming.dstream.DStream.createRDDWithLocalProperties(DStream.scala:426)
> at org.apache.spark.streaming.dstream.ForEachDStream$$anonfun$1.apply$mcV$sp(ForEachDStream.scala:49)
> at org.apache.spark.streaming.dstream.ForEachDStream$$anonfun$1.apply(ForEachDStream.scala:49)
> at org.apache.spark.streaming.dstream.ForEachDStream$$anonfun$1.apply(ForEachDStream.scala:49)
> at scala.util.Try$.apply(Try.scala:161)
> at org.apache.spark.streaming.scheduler.Job.run(Job.scala:39)^C
> at org.apache.spark.streaming.scheduler.JobScheduler$JobHandler$$anonfun$run$1.apply$mcV$sp(JobScheduler.scala:224)
> at org.apache.spark.streaming.scheduler.JobScheduler$JobHandler$$anonfun$run$1.apply(JobScheduler.scala:224)
> at org.apache.spark.streaming.scheduler.JobScheduler$JobHandler$$anonfun$run$1.apply(JobScheduler.scala:224)
> at scala.util.DynamicVariable.withValue(DynamicVariable.scala:57)
> at org.apache.spark.streaming.scheduler.JobScheduler$JobHandler.run(JobScheduler.scala:223)
> at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
> at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
> at java.lang.Thread.run(Thread.java:745)
> Caused by: org.kie.scanner.embedder.MavenEmbedderException: com.google.inject.ProvisionException: Guice provision errors:
> 1) No implementation for org.apache.maven.repository.RepositorySystem was bound.
> while locating org.apache.maven.execution.DefaultMavenExecutionRequestPopulator
> at ClassRealm[maven, parent: ClassRealm[maven-parent, parent: null]]
> at ClassRealm[maven, parent: ClassRealm[maven-parent, parent: null]]
> while locating org.apache.maven.execution.MavenExecutionRequestPopulator
> 1 error
> role: org.apache.maven.execution.MavenExecutionRequestPopulator
> roleHint:
> at org.kie.scanner.embedder.MavenEmbedder.<init>(MavenEmbedder.java:94)
> at org.kie.scanner.embedder.MavenEmbedder.<init>(MavenEmbedder.java:75)
> at org.kie.scanner.embedder.MavenEmbedder.<init>(MavenEmbedder.java:69)
> at org.kie.scanner.embedder.MavenProjectLoader.parseMavenPom(MavenProjectLoader.java:55)
> ... 28 more
> Caused by: org.codehaus.plexus.component.repository.exception.ComponentLookupException: com.google.inject.ProvisionException: Guice provision errors:
> 1) No implementation for org.apache.maven.repository.RepositorySystem was bound.
> while locating org.apache.maven.execution.DefaultMavenExecutionRequestPopulator
> at ClassRealm[maven, parent: ClassRealm[maven-parent, parent: null]]
> at ClassRealm[maven, parent: ClassRealm[maven-parent, parent: null]]
> while locating org.apache.maven.execution.MavenExecutionRequestPopulator
> 1 error
> role: org.apache.maven.execution.MavenExecutionRequestPopulator
> roleHint:
> at org.codehaus.plexus.DefaultPlexusContainer.lookup(DefaultPlexusContainer.java:264)
> at org.codehaus.plexus.DefaultPlexusContainer.lookup(DefaultPlexusContainer.java:252)
> at org.codehaus.plexus.DefaultPlexusContainer.lookup(DefaultPlexusContainer.java:246)
> at org.kie.scanner.embedder.PlexusComponentProvider.lookup(PlexusComponentProvider.java:42)
> at org.kie.scanner.embedder.MavenEmbedder.buildMavenExecutionRequest(MavenEmbedder.java:111)
> at org.kie.scanner.embedder.MavenEmbedder.<init>(MavenEmbedder.java:84)
> ... 31 more
> Caused by: com.google.inject.ProvisionException: Guice provision errors:
> 1) No implementation for org.apache.maven.repository.RepositorySystem was bound.
> while locating org.apache.maven.execution.DefaultMavenExecutionRequestPopulator
> at ClassRealm[maven, parent: ClassRealm[maven-parent, parent: null]]
> at ClassRealm[maven, parent: ClassRealm[maven-parent, parent: null]]
> while locating org.apache.maven.execution.MavenExecutionRequestPopulator
> 1 error
> at com.google.inject.internal.InjectorImpl$3.get(InjectorImpl.java:974)
> at org.eclipse.sisu.inject.LazyBeanEntry.getValue(LazyBeanEntry.java:82)
> at org.eclipse.sisu.plexus.LazyPlexusBean.getValue(LazyPlexusBean.java:51)
> at org.codehaus.plexus.DefaultPlexusContainer.lookup(DefaultPlexusContainer.java:260)
> ... 36 more
> {code}
--
This message was sent by Atlassian JIRA
(v6.4.11#64026)
8 years, 7 months
[JBoss JIRA] (WFLY-6578) Add driver module slot configuration to Datasource subsystem
by Mathieu Lachance (JIRA)
[ https://issues.jboss.org/browse/WFLY-6578?page=com.atlassian.jira.plugin.... ]
Mathieu Lachance commented on WFLY-6578:
----------------------------------------
https://github.com/wildfly/wildfly/pull/8899/commits
> Add driver module slot configuration to Datasource subsystem
> ------------------------------------------------------------
>
> Key: WFLY-6578
> URL: https://issues.jboss.org/browse/WFLY-6578
> Project: WildFly
> Issue Type: Feature Request
> Components: JCA
> Affects Versions: 10.0.0.Final
> Reporter: Mathieu Lachance
> Assignee: Jesper Pedersen
>
> Currently, it is possible to define Datasource and Drivers as such in the standalone.xml:
> {code}
> <subsystem xmlns="urn:jboss:domain:datasources:4.0">
> <datasources>
> <datasource jndi-name="java:jboss/datasources/ExampleDS" pool-name="ExampleDS" enabled="true" use-java-context="true">
> <connection-url>jdbc:h2:mem:test;DB_CLOSE_DELAY=-1;DB_CLOSE_ON_EXIT=FALSE</connection-url>
> <driver>h2</driver>
> <security>
> <user-name>sa</user-name>
> <password>sa</password>
> </security>
> </datasource>
> <drivers>
> <driver name="h2" module="com.h2database.h2">
> <xa-datasource-class>org.h2.jdbcx.JdbcDataSource</xa-datasource-class>
> </driver>
> </drivers>
> </datasources>
> </subsystem>
> {code}
> The driver module attribute allows to point to a well define module which is fine. Though it doesn't seems possible to point to another "slot" to be able to switch between multiple version of the same driver.
> It would be nice if we could do so, ex:
> {code}
> <driver name="h2" module="com.h2database.h2" module-slot="1.0">
> <xa-datasource-class>org.h2.jdbcx.JdbcDataSource</xa-datasource-class>
> </driver>
> {code}
> Also, while playing with the configuration, I've noticed that whatever you put onto the driver definition will be safely discarded instead of throwing an XSD validation exception. I'm not sure this is the proper behavior as it could be very misleading, ex:
> {code}
> <driver name="h2" module="com.h2database.h2" module-slot="1.0">
> {code}
> Currenly doesn't throw any exception and is getting rewrited to:
> {code}
> <driver name="h2" module="com.h2database.h2">
> {code}
> Should a bug be opened for that? To me it looks a bit similar to: WFLY-4464
> The only workaround I'm aware of is to either:
> 1. create a new main module.xml that depends on the actual versioned slot
> 2. have the versioning scheme part of the dependency (ex: com/h2database-1.0/main)
> Both solutions are not very elegant.
> Thanks,
--
This message was sent by Atlassian JIRA
(v6.4.11#64026)
8 years, 7 months
[JBoss JIRA] (WFLY-6578) Add driver module slot configuration to Datasource subsystem
by Jesper Pedersen (JIRA)
[ https://issues.jboss.org/browse/WFLY-6578?page=com.atlassian.jira.plugin.... ]
Jesper Pedersen commented on WFLY-6578:
---------------------------------------
Those should make good contributions, so feel free to open JIRAs and assign them to yourself.
> Add driver module slot configuration to Datasource subsystem
> ------------------------------------------------------------
>
> Key: WFLY-6578
> URL: https://issues.jboss.org/browse/WFLY-6578
> Project: WildFly
> Issue Type: Feature Request
> Components: JCA
> Affects Versions: 10.0.0.Final
> Reporter: Mathieu Lachance
> Assignee: Jesper Pedersen
>
> Currently, it is possible to define Datasource and Drivers as such in the standalone.xml:
> {code}
> <subsystem xmlns="urn:jboss:domain:datasources:4.0">
> <datasources>
> <datasource jndi-name="java:jboss/datasources/ExampleDS" pool-name="ExampleDS" enabled="true" use-java-context="true">
> <connection-url>jdbc:h2:mem:test;DB_CLOSE_DELAY=-1;DB_CLOSE_ON_EXIT=FALSE</connection-url>
> <driver>h2</driver>
> <security>
> <user-name>sa</user-name>
> <password>sa</password>
> </security>
> </datasource>
> <drivers>
> <driver name="h2" module="com.h2database.h2">
> <xa-datasource-class>org.h2.jdbcx.JdbcDataSource</xa-datasource-class>
> </driver>
> </drivers>
> </datasources>
> </subsystem>
> {code}
> The driver module attribute allows to point to a well define module which is fine. Though it doesn't seems possible to point to another "slot" to be able to switch between multiple version of the same driver.
> It would be nice if we could do so, ex:
> {code}
> <driver name="h2" module="com.h2database.h2" module-slot="1.0">
> <xa-datasource-class>org.h2.jdbcx.JdbcDataSource</xa-datasource-class>
> </driver>
> {code}
> Also, while playing with the configuration, I've noticed that whatever you put onto the driver definition will be safely discarded instead of throwing an XSD validation exception. I'm not sure this is the proper behavior as it could be very misleading, ex:
> {code}
> <driver name="h2" module="com.h2database.h2" module-slot="1.0">
> {code}
> Currenly doesn't throw any exception and is getting rewrited to:
> {code}
> <driver name="h2" module="com.h2database.h2">
> {code}
> Should a bug be opened for that? To me it looks a bit similar to: WFLY-4464
> The only workaround I'm aware of is to either:
> 1. create a new main module.xml that depends on the actual versioned slot
> 2. have the versioning scheme part of the dependency (ex: com/h2database-1.0/main)
> Both solutions are not very elegant.
> Thanks,
--
This message was sent by Atlassian JIRA
(v6.4.11#64026)
8 years, 7 months
[JBoss JIRA] (WFCORE-1503) Reading the logging subsystem resource can be slow if several log files exist
by James Perkins (JIRA)
[ https://issues.jboss.org/browse/WFCORE-1503?page=com.atlassian.jira.plugi... ]
James Perkins updated WFCORE-1503:
----------------------------------
Labels: 2.2 (was: )
> Reading the logging subsystem resource can be slow if several log files exist
> -----------------------------------------------------------------------------
>
> Key: WFCORE-1503
> URL: https://issues.jboss.org/browse/WFCORE-1503
> Project: WildFly Core
> Issue Type: Bug
> Components: Logging
> Reporter: James Perkins
> Assignee: James Perkins
> Labels: 2.2
>
> The logging subsystem uses a dynamic resource to list log files as resources. These are the files that are allowed to be read or downloaded and are attached to known file handlers.
> When several file handlers are defined the file tree walker is invoked several times per {{read-resource-operation}}. This may be happening for each stage of an operation. Ideally we could either cache the file listing per-operation or only execute at the runtime stage. This may not be an option for dynamic resources however.
> As a result of this poor performance the web console is very slow to load the logging subsystem configuration. This is not an issue with the web console.
--
This message was sent by Atlassian JIRA
(v6.4.11#64026)
8 years, 7 months
[JBoss JIRA] (WFLY-6577) Unable to build Hibernate SessionFactory
by Scott Marlow (JIRA)
[ https://issues.jboss.org/browse/WFLY-6577?page=com.atlassian.jira.plugin.... ]
Scott Marlow commented on WFLY-6577:
------------------------------------
Thank you, I'll close this WFLY-6577 as a duplicate of HHH-10719.
> Unable to build Hibernate SessionFactory
> ----------------------------------------
>
> Key: WFLY-6577
> URL: https://issues.jboss.org/browse/WFLY-6577
> Project: WildFly
> Issue Type: Bug
> Components: JPA / Hibernate
> Affects Versions: 10.0.0.Final
> Environment: Windows 7 Professional (64-bit)
> Java 8 u91 (64-bit)
> PostgreSQL 9.5 on Windows (64-bit)
> Reporter: vbndeveloper
> Assignee: Scott Marlow
> Original Estimate: 1 hour
> Remaining Estimate: 1 hour
>
> The hibernate entity to schema validator fails on a scenario where you have a bean with a nullable Boolean getter/setter and also a convenience primitive boolean method decorated with a @Transient annotation.
> Ideally the validator would recognize javax.persistence.Transient annotated methods and ignore them when validating the database schema.
--
This message was sent by Atlassian JIRA
(v6.4.11#64026)
8 years, 7 months