Quantcast
Channel: Pentaho Community Forums
Viewing all 16689 articles
Browse latest View live

第5回 オープンソースbi勉強会


Want to access user based data using xaction file in penaho 5.02...

$
0
0
Hi,

I am new to pentaho. I have a requirement to show user based report in pentaho. I have searched and came to know that this is possible using xaction files.
I am using Pentaho design studio for creating xaction file, but my requirement is of pentaho 5.02 and xaction's are not supported in this pentaho version...

Is there any way to use xaction files in pentaho 5.02...


Thanks in advance.

how to converti nominal to numeric

$
0
0
Hi all ,
I have a dataset with nominal and numeric attributes. I want to convert the nominal attributes to numeric. I used Nominaltobinay filter for converting my nominal attribute which has 53 distinct values and this resulted in 53 new columns in the dataset " for only one attribute"" .


Is there a better way of representing my data or converting the nominal to numeric type? I want to represent the nominal attribute with a single numeric value.


Note: dataset was csv but when opened with weka numerics are change to nominal
and i am try to change " the arff file manually using notepad" the head of attribute to numeric but not works

thanks for any help

PDI 5.0 kettle reading excel sheet and csv file

$
0
0
Hi,

I have a PDI 5.0 kettle scenario where...

1. I have a csv Input file with 200 columns.
2. I have an MS Excel workbook with 10 sheets. In each sheet I will specify only 10 - 15 column names (10-15 column names alone) that are already present in the csv file mentioned above. The column names are same both in csv and excel.
3. I need to select columns from the csv file that are already specified in the excel sheet and write it to a separate text files.

4. I need to repeat this step for all sheets present in that excel file. (Read the column names from the excel sheet, Select them from the csv file and writing it to a separate text files)

Please help me achieve this. I am very new to pentaho.
Its very very important for me.

Thanks
Dhan

LEVEL_NUMBER in Mdx QUery

$
0
0
i'm using an odbo driver implementation ( https://sourceforge.net/projects/xmlaconnect/ by Alexandra Iancu ) to use excel pivot table with mondrian .
i have tried to use the function "Insert Slicer" but the mdx query generated from excel is wrong for Mondrian .

with set __XLUniqueNames as { [CUSTOMER_D.CUSTOMER_CLASSECF1_H].[Tutte le Categorie Cliente] }
set __XLDrilledUp as
Generate(__XLUniqueNames,
{ IIF([CUSTOMER_D.CUSTOMER_CLASSECF1_H].currentmember.LEVEL_NUMBER <= 2,
[CUSTOMER_D.CUSTOMER_CLASSECF1_H].currentmember,
Ancestor([CUSTOMER_D.CUSTOMER_CLASSECF1_H].currentmember,
[CUSTOMER_D.CUSTOMER_CLASSECF1_H].currentmember.LEVEL_NUMBER - 2)) } )
member [Measures].__XLPath as
Generate(
Ascendants([CUSTOMER_D.CUSTOMER_CLASSECF1_H].currentmember),
[CUSTOMER_D.CUSTOMER_CLASSECF1_H].currentmember.unique_name,
"__XLPSEP")
select { [Measures].__XLPath } on 0,
__XLDrilledUp on 1
from [FATTURATO]
cell properties value

The error returned from mondrian was :

The Mondrian XML: Mondrian Error:No function matches signature Member.LEVEL_NUMBER;

I consulted the documentation of mondrian and the property exists ( http://mondrian.pentaho.com/api/mond...l#LEVEL_NUMBER )

Is there any way to modify this query so that it work ( in the xmla connector i can define alias of function so i can adapt the mdx query to mondrian ) .
Alternatively how i should modify the mondrian code to implement this functionality .
Thanks for any answer .








Pentaho 5.0.1 ESAPI.properties problem

$
0
0
When I run my BI Server on Windows platform all working but when I copy same files on Linux Server Login page successfully loads but than server die.

My catalina.out log file on linux platform:

Code:

INFO: Server startup in 18706 ms
Attempting to load ESAPI.properties via file I/O.
Attempting to load ESAPI.properties as resource file via file I/O.
Not found in 'org.owasp.esapi.resources' directory or file not readable: /usr/pentaho/biserver-ce/tomcat/bin/ESAPI.properties
Not found in SystemResource Directory/resourceDirectory: .esapi/ESAPI.properties
Not found in 'user.home' (/root) directory: /root/esapi/ESAPI.properties
Loading ESAPI.properties via file I/O failed. Exception was: java.io.FileNotFoundException
Attempting to load ESAPI.properties via the classpath.
SUCCESSFULLY LOADED ESAPI.properties via the CLASSPATH from '/ (root)' using current thread context class loader!
SecurityConfiguration for Validator.ConfigurationFile not found in ESAPI.properties. Using default: validation.properties
Attempting to load validation.properties via file I/O.
Attempting to load validation.properties as resource file via file I/O.
Not found in 'org.owasp.esapi.resources' directory or file not readable: /usr/pentaho/biserver-ce/tomcat/bin/validation.properties
Not found in SystemResource Directory/resourceDirectory: .esapi/validation.properties
Not found in 'user.home' (/root) directory: /root/esapi/validation.properties
Loading validation.properties via file I/O failed.
Attempting to load validation.properties via the classpath.
validation.properties could not be loaded by any means. fail. Exception was: java.lang.IllegalArgumentException: Failed to load ESAPI.properties as a classloader resource.

After above actions Login page load and existing users can`t log in!

Parameter caching help needed

$
0
0
Hi All,

We have a few parameterized inventory and sales performance reports. The parameters are typically text boxes to lookup values such as brand names, part numbers, etc. Certain parameters are linked. For e.g selecting a brand name retrieves only part numbers related to that brand. Each parameter can have 10k to 1 million unique values.

The problem we are facing is that the parameter values are blank during initial loading, due to which the reporting engine fetches all values for all parameters, causing the report to crash. We tried reducing the eh-cache size; however, the cache remains static and only a fixed set of values are shown in the dropdown (for e.g. first 100 brand names).

Due to this, we had to disable caching for the reports.

Is there any way we can:
1. Ask the reporting engine not to load all parameter values in memory on-load
2. Load only certain parameters in memory on report load. For e.g. load only brand names, not part numbers
3. If a parameter value is not in the cache, force the reporting engine to look in the DB. This would be useful for linked parameters. For e.g. load only part numbers related to the brand selected in the brand name lookup
4. Disable the auto-lookup feature (as these values are always retrieved from cache only)

Please help.

Thanks,
Vinod

Error previewing an excel file when all the fields are empty - Pentaho 5.0.3

$
0
0
Hi All,

When I tried to preview the Xlsx file data using Microsoft Excel Input step, It is throwing the following error :

ERROR (version 5.0.1, build 1 from 2013-10-30_19-53-32 by buildguy) : Error processing row from Excel file [D:\sample.xlsx] :
org.apache.poi.openxml4j.exceptions.OpenXML4JRuntimeException: Fail to save: an error occurs while saving the package : null
2014/04/20 20:42:13 - Microsoft Excel Input.0 - ERROR (version 5.0.1, build 1 from 2013-10-30_19-53-32 by buildguy) : org.apache.poi.openxml4j.exceptions.OpenXML4JRuntimeException: Fail to save: an error occurs while saving the package : null
2014/04/20 20:42:13 - Microsoft Excel Input.0 - at org.apache.poi.openxml4j.opc.ZipPackage.saveImpl(ZipPackage.java:500)
2014/04/20 20:42:13 - Microsoft Excel Input.0 - at org.apache.poi.openxml4j.opc.OPCPackage.save(OPCPackage.java:1417)
2014/04/20 20:42:13 - Microsoft Excel Input.0 - at org.apache.poi.openxml4j.opc.OPCPackage.save(OPCPackage.java:1404)
2014/04/20 20:42:13 - Microsoft Excel Input.0 - at org.apache.poi.openxml4j.opc.ZipPackage.closeImpl(ZipPackage.java:349)
2014/04/20 20:42:13 - Microsoft Excel Input.0 - at org.apache.poi.openxml4j.opc.OPCPackage.close(OPCPackage.java:420)
2014/04/20 20:42:13 - Microsoft Excel Input.0 - at org.pentaho.di.trans.steps.excelinput.poi.PoiWorkbook.close(PoiWorkbook.java:101)
2014/04/20 20:42:13 - Microsoft Excel Input.0 - at org.pentaho.di.trans.steps.excelinput.ExcelInput.jumpToNextFile(ExcelInput.java:757)
2014/04/20 20:42:13 - Microsoft Excel Input.0 - at org.pentaho.di.trans.steps.excelinput.ExcelInput.getRowFromWorkbooks(ExcelInput.java:717)
2014/04/20 20:42:13 - Microsoft Excel Input.0 - at org.pentaho.di.trans.steps.excelinput.ExcelInput.processRow(ExcelInput.java:471)
2014/04/20 20:42:13 - Microsoft Excel Input.0 - at org.pentaho.di.trans.step.RunThread.run(RunThread.java:60)
2014/04/20 20:42:13 - Microsoft Excel Input.0 - at java.lang.Thread.run(Unknown Source)
2014/04/20 20:42:13 - Microsoft Excel Input.0 - Caused by: java.lang.NullPointerException
2014/04/20 20:42:13 - Microsoft Excel Input.0 - at org.apache.poi.openxml4j.opc.internal.marshallers.ZipPackagePropertiesMarshaller.marshall(ZipPackagePropertiesMarshaller.java:46)
2014/04/20 20:42:13 - Microsoft Excel Input.0 - at org.apache.poi.openxml4j.opc.ZipPackage.saveImpl(ZipPackage.java:444)
2014/04/20 20:42:13 - Microsoft Excel Input.0 - ... 10 more


It is able to read the data, till any of the fields contains the data, If all the fields are empty then it is throwing the above mentioned error.


Note: I'm using Excel 2007 XLSX (Apache POI) as Spread Sheet type (engine).



Any help is highly appreciated.


Thanks,
Siva

Dashboard: import dynamic CSV data

$
0
0
Hello everybody. First sorry for my English!

I have designed a dashboard with CDE inside Pentaho Business Analytics. To show de data in a chart, I use a CSV with the "Data Source Wizard".
Now I have an application running in my PC locally and it generates a different CSV file in each execution. I would like the dashboard shows the newest data, but I don't want to load the CSV into Pentaho BA every time it changes. Is there any solution to this? :eek:

Thanks in advance! :)
Regards.

How to Improve performance of table output

$
0
0
Hello,

I am fetching the millians of records from source to Target
Source -- oracle DB 12.1.3
Target -- MySQL 5.5

I am using table input step and table output step.
input/output steps inserting 3 rows per sec.
Can anybody give me suggestions, how to improve performance of input/output steps.

Thankyou,
Satya

Dashboard: insert checkbox into table

$
0
0
Hello! First sorry for my English.

I am making a dashboard with CDE inside Pentaho Business Analytics. I have inserted a table component with several rows. I wonder if it is possible to add a column with a check component (a.k.a. checkbox) for each table row.

Thanks in advance. :)
Regards.

Text File Output format field question

$
0
0
Hi all.

I have simple text file output example.
Can please someone explain me how does Length/Precision/Currency/Decimal/Group settings can communicate with each other?
I don't want to use 'Format' field - just using other to check it is works properly.

Just for testing purposes I have input as a BigDecimal like

1300025525.002356898

...and for file output step I want to see what happens if:

this is file output settings
fileOutSettings.jpg

this is file output
fileOut.jpg

So questions:
1) Why is $ is not shown?
2) Why grouping symbol not shown?
3) Why decimal is truncated to one symbol?
4) Is Length is taken into account?

Once again this is only test-functionality example - to see how this field output settings can communicate each other.

Thank you in advance.

This is simple transformation.
fileOutput.ktr
Attached Images
Attached Files

CCC2 - Text Angle for Dot Chart

$
0
0
Hello.

I'm looking for the way to put text angle in a dot chart component for CE 4.5. With CCC I worked with xAxisLabel_textAngle=-0.3, but this extension point doesn't work with CCC2 (2014.03.07). I've tried unsuccessfully with baseAxisLabel_textAngle=0.3 or -Math.PI/6.

Anyone know how to do?

Thanks!

Convert Binary Data from Cassandra DB to String/Int

$
0
0
Hello. I am using Cassandra Input as the input in the Kettle. I have defined the required connections etc in the input.

For output, in the column value I am getting the value as Binary (Machine Character). The value is stored in Cassandra from application end using ObjectSerializer.Hence the column value is always Binary.

I need to convert this binary value to string/int, once I have the value from the execution of CQL.

I used the Modified Java Script step after the Cassandra Input and used.

var msg=new java.lang.String(columnValue,"UTF-8");

But, the output column "msg" prints machine character.

Is there any way, I can convert the binary data to different data type such as string/int etc.

Found good video tutorial on Pentaho !


Got permission issue while un-installing pentaho cdf from marketplace

$
0
0
Hi All,


Got permission issue while un-installing pentaho cdf . Can some one help me to know where can i changes role setting to unistall ctools plugin .

Error uninstalling
Unauthorized Access. Your Pentaho roles do not allow you to make changes to plugins.




Thanks

Saving Hierarchical data in Cassandra

$
0
0
Hi


I have a Time Series Data that I want to persist in Cassandra database.


Example Sheet : Here t1, t2, t3 are my time series values while a,b,c,d,e,f are my hierarchies.


Composite_column Value
a:b:c:t1 v1
a:b:c:t2 v4
a:b:d:t2 v2
a:e:f:t3 v3


Now, for each row I want to create a composite column for the rows under the field Composite_column and add to it the value under the field Value. The default action of cassandra output transformation in Pentaho only allows for creation of 3 separate rows with keys in the field Composite_column and values in the field Value. How do I get the cassandra output step to perform the desired behavior ie. persisting all these entries in a single row of cassandra database?

hello

Pentaho Report Bursting with KETTLE, struggling with PARAMETERS.

$
0
0
Hi, I am trying to do report bursting through kettle. But i am facing some problems while trying to pass some parameter values from kettle to PRD (using pentaho reporting output step).here is the following steps i am doing.data grid --> add constants --> modified java scrip value --> pentaho reporting output --> mailabove flow works fine with no parameters specified in pentaho reporting output step. It simply calls the .prpt file and generates .pdf file(with whole data in each file as there are no parameters) and send mails.But when i put a parameter in my .prpt file i.e "select * from table where date = ${memberdate}" and try to run my transformation it gives me something shown in the attachment.Any help would be great.http://forums.pentaho.com/images/smilies/confused.pngThanksPrabal
Attached Files

Getting Header data for a Excel File with dynamic fields

$
0
0
Hi

I have a tabular data in excel which has a header.

The header length & its names are not constants and change with time.

Is there a way that Excel input can dynamically find the header information when we had defined a transformation without using "Get Fields" from the Excel Input?
Viewing all 16689 articles
Browse latest View live