This PHP library offers a way to read data from, and write data to, a range of file formats and media. Additionally, it includes tools to manipulate your data.
- Read from and write to CSV files, Excel files, databases, and more.
- Convert between charsets, dates, strings and objects on the fly.
- Build reusable and extensible import workflows.
- Decoupled components that you can use on their own, such as a CSV and Excel reader and writer.
- Well-tested code.
- Installation
- Usage
- Running the tests
- License
This library is available on Packagist. The recommended way to install it is through Composer:
$ composer require ddeboer/data-import:@stable
Then include Composer’s autoloader:
require_once 'vendor/autoload.php';
For integration with Symfony2 projects, the DdeboerDataImportBundle is available.
Broadly speaking, you can use this library in two ways:
- organize your import process around a workflow, or
- use one or more of the components on their own, such as readers, writers or converters.
Each data import revolves around the workflow and takes place along the following lines:
- Construct a reader.
- Construct a workflow and pass the reader to it, optionally pass a logger as second argument. Add at least one writer to the workflow.
- Optionally, add filters, item converters and value converters to the workflow.
- Process the workflow. This will read the data from the reader, filter and
convert the data, and write the output to each of the writers. The process method also
returns a
Result
object which contains various information about the import.
In other words, the workflow acts as a mediator between a reader and one or more writers, filters and converters.
Optionally you can skip items on failure like this $workflow->setSkipItemOnFailure(true)
.
Errors will be logged if you have passed a logger to the workflow constructor.
Schematically:
use Ddeboer\DataImport\Workflow;
use Ddeboer\DataImport\Reader;
use Ddeboer\DataImport\Writer;
use Ddeboer\DataImport\Filter;
$reader = new Reader\...;
$workflow = new Workflow($reader, $logger);
$result = $workflow
->addWriter(new Writer\...())
->addWriter(new Writer\...())
->addFilter(new Filter\CallbackFilter(...))
->setSkipItemOnFailure(true)
->process()
;
The Workflow Result object exposes various methods which you can use to decide what to do after an import.
The result will be an instance of Ddeboer\DataImport\Result
. It is automatically created and populated by the
Workflow
. It will be returned to you after calling the process()
method on the Workflow
The Result
provides the following methods:
//the name of the import - which is an optional 3rd parameter to
//the Workflow class. Returns null by default.
public function getName();
//DateTime instance created at the start of the import.
public function getStartTime();
//DateTime instance created at the end of the import.
public function getEndTime();
//DateInterval instance. Diff off the start + end times.
public function getElapsed();
//Count of exceptions which caught by the Workflow.
public function getErrorCount();
//Count of processed items minus the count of exceptions caught.
public function getSuccessCount();
//Count of items processed
//This will not include any filtered items or items which fail conversion.
public function getTotalProcessedCount();
//bool to indicate whether any exceptions were caught.
public function hasErrors();
//An array of exceptions caught by the Workflow.
public function getExceptions();
Example use cases:
- You want to send an e-mail with the results of the import
- You want to send a Text alert if a particular file failed
- You want to move an import file to a failed directory if there were errors
- You want to log how long imports are taking
Readers read data that will be imported by iterating over it. This library includes a handful of readers. Additionally, you can easily implement your own.
You can use readers on their own, or construct a workflow from them:
$workflow = new Workflow($reader);
Reads arrays. Most useful for testing your workflow.
Reads CSV files, and is optimized to use as little memory as possible.
use Ddeboer\DataImport\Reader\CsvReader;
$file = new \SplFileObject('/path/to/csv_file.csv');
$reader = new CsvReader($file);
Optionally construct with different delimiter, enclosure and/or escape character:
$reader = new CsvReader($file, ';');
Then iterate over the CSV file:
foreach ($reader as $row) {
// $row will be an array containing the comma-separated elements of the line:
// array(
// 0 => 'James',
// 1 => 'Bond'
// etc...
// )
}
If one of your rows contains column headers, you can read them to make the rows associative arrays:
$reader->setHeaderRowNumber(0);
foreach ($reader as $row) {
// $row will now be an associative array:
// array(
// 'firstName' => 'James',
// 'lastName' => 'Bond'
// etc...
// )
}
The CSV reader operates in strict mode by default. If the reader encounters a
row where the number of values differs from the number of column headers, an
error is logged and the row is skipped. Retrieve the errors with getErrors()
.
To disable strict mode, set $reader->setStrict(false)
after you instantiate
the reader.
Disabling strict mode means:
- Any rows that contain fewer values than the column headers are simply padded with null values.
- Any additional values in a row that contain more values than the column headers are ignored.
Examples where this is useful:
- Outlook 2010: which omits trailing blank values
- Google Contacts: which exports more values than there are column headers
Sometimes a CSV file contains duplicate column headers, for instance:
id | details | details |
---|---|---|
1 | bla | more bla |
By default, a DuplicateHeadersException
will be thrown if you call
setHeaderRowNumber(0)
on this file. You can handle duplicate columns in
one of three ways:
- call
setColumnHeaders(['id', 'details', 'details_2'])
to specify your own headers - call
setHeaderRowNumber
with theCsvReader::DUPLICATE_HEADERS_INCREMENT
flag to generate incremented headers; in this case:id
,details
anddetails1
- call
setHeaderRowNumber
with theCsvReader::DUPLICATE_HEADERS_MERGE
flag to merge duplicate values into arrays; in this case, the first row’s values will become:[ 'id' => 1, 'details' => [ 'bla', 'more bla' ] ]
.
Reads data through Doctrine’s DBAL. Your project should include Doctrine’s DBAL package:
$ composer require doctrine/dbal
use Ddeboer\DataImport\Reader\DbalReader;
$reader = new DbalReader(
$connection, // Instance of \Doctrine\DBAL\Connection
'SELECT u.id, u.username, g.name FROM `user` u INNER JOIN groups g ON u.group_id = g.id'
);
Reads data through the Doctrine ORM:
use Ddeboer\DataImport\Reader\DoctrineReader;
$reader = new DoctrineReader($entityManager, 'Your\Namespace\Entity\User');
Acts as an adapter for the PHPExcel library. Make sure to include that library in your project:
$ composer require phpoffice/phpexcel
Then use the reader to open an Excel file:
use Ddeboer\DataImport\Reader\ExcelReader;
$file = new \SplFileObject('path/to/ecxel_file.xls');
$reader = new ExcelReader($file);
To set the row number that headers will be read from, pass a number as the second argument.
$reader = new ExcelReader($file, 2);
To read the specific sheet:
$reader = new ExcelReader($file, null, 3);
###OneToManyReader
Allows for merging of two data sources (using existing readers), for example you have one CSV with orders and another with order items.
Imagine two CSV's like the following:
OrderId,Price
1,30
2,15
OrderId,Name
1,"Super Cool Item 1"
1,"Super Cool Item 2"
2,"Super Cool Item 3"
You want to associate the items to the order. Using the OneToMany reader we can nest these rows in the order using a key which you specify in the OneToManyReader.
The code would look something like:
$orderFile = new \SplFileObject("orders.csv");
$orderReader = new CsvReader($file, $orderFile);
$orderReader->setHeaderRowNumber(0);
$orderItemFile = new \SplFileObject("order_items.csv");
$orderItemReader = new CsvReader($file, $orderFile);
$orderItemReader->setHeaderRowNumber(0);
$oneToManyReader = new OneToManyReader($orderReader, $orderItemReader, 'items', 'OrderId', 'OrderId');
The third parameter is the key which the order item data will be nested under. This will be an array of order items. The fourth and fifth parameters are "primary" and "foreign" keys of the data. The OneToMany reader will try to match the data using these keys. Take for example the CSV's given above, you would expect that Order "1" has the first 2 Order Items associated to it due to their Order Id's also being "1".
Note: You can omit the last parameter, if both files have the same field. Eg if parameter 4 is 'OrderId' and you don't specify parameter 5, the reader will look for the foreign key using 'OrderId'
The resulting data will look like:
//Row 1
array(
'OrderId' => 1,
'Price' => 30,
'items' => array(
array(
'OrderId' => 1,
'Name' => 'Super Cool Item 1',
),
array(
'OrderId' => 1,
'Name' => 'Super Cool Item 2',
),
),
);
//Row2
array(
'OrderId' => 2,
'Price' => 15,
'items' => array(
array(
'OrderId' => 2,
'Name' => 'Super Cool Item 1',
),
)
);
You can create your own data reader by implementing the ReaderInterface.
Resembles the ArrayReader. Probably most useful for testing your workflow.
Writes CSV files:
use Ddeboer\DataImport\Writer\CsvWriter;
$writer = new CsvWriter();
$writer->setStream(fopen('output.csv', 'w'));
// Write column headers:
$writer->writeItem(array('first', 'last'));
$writer
->writeItem(array('James', 'Bond'))
->writeItem(array('Auric', 'Goldfinger'))
->finish();
Writes data through Doctrine:
use Ddeboer\DataImport\Writer\DoctrineWriter;
$writer = new DoctrineWriter($entityManager, 'YourNamespace:Employee');
$writer
->prepare()
->writeItem(
array(
'first' => 'James',
'last' => 'Bond'
)
)
->finish();
By default, DoctrineWriter will truncate your data before running the workflow.
Call disableTruncate()
if you don't want this.
If you are not truncating data, DoctrineWriter will try to find an entity having it's primary key set to the value of the first column of the item. If it finds one, the entity will be updated, otherwise it's inserted. You can tell DoctrineWriter to lookup the entity using different columns of your item by passing a third parameter to it's constructor.
$writer = new DoctrineWriter($entityManager, 'YourNamespace:Employee', 'columnName');
or
$writer = new DoctrineWriter($entityManager, 'YourNamespace:Employee', array('column1', 'column2', 'column3'));
The DoctrineWriter will also search out associations automatically and link them by an entity reference. For example suppose you have a Product entity that you are importing and must be associated to a Category. If there is a field in the import file named 'Category' with an id, the writer will use metadata to get the association class and create a reference so that it can be associated properly. The DoctrineWriter will skip any association fields that are already objects in cases where a converter was used to retrieve the association.
Use the PDO writer for importing data into a relational database (such as MySQL, SQLite or MS SQL) without using Doctrine.
use Ddeboer\DataImport\Writer\PdoWriter;
$pdo = new \PDO('sqlite::memory:');
$pdo->setAttribute(\PDO::ATTR_ERRMODE, \PDO::ERRMODE_EXCEPTION);
$writer = new PdoWriter($pdo, 'my_table');
Writes data to an Excel file. It requires the PHPExcel package:
$ composer require phpoffice/phpexcel
use Ddeboer\DataImport\Writer\ExcelWriter;
$file = new \SplFileObject('data.xlsx', 'w');
$writer = new ExcelWriter($file);
$writer
->prepare()
->writeItem(array('first', 'last'))
->writeItem(array('first' => 'James', 'last' => 'Bond'))
->finish();
You can specify the name of the sheet to write to:
$writer = new ExcelWriter($file, 'My sheet');
You can open an already existing file and add a sheet to it:
$file = new \SplFileObject('data.xlsx', 'a'); // Open file with append mode
$writer = new ExcelWriter($file, 'New sheet');
If you wish to overwrite an existing sheet instead, specify the name of the existing sheet:
$writer = new ExcelWriter($file, 'Old sheet');
This writer displays items as table on console output for debug purposes when you start the workflow from the command-line. It requires Symfony’s Console component 2.5 or higher:
$ composer require symfony/console ~2.5
use Ddeboer\DataImport\Reader;
use Ddeboer\DataImport\Writer\ConsoleTableWriter;
use Symfony\Component\Console\Output\ConsoleOutput;
use Symfony\Component\Console\Helper\Table;
$reader = new Reader\...;
$output = new ConsoleOutput(...);
$table = new Table($output);
// Make some manipulations, e.g. set table style
$table->setStyle('compact');
$workflow = new Workflow($reader);
$workflow->addWriter(new ConsoleTableWriter($output, $table));
This writer displays import progress when you start the workflow from the command-line. It requires Symfony’s Console component:
$ composer require symfony/console
use Ddeboer\DataImport\Writer\ConsoleProgressWriter;
use Symfony\Component\Console\Output\ConsoleOutput;
$output = new ConsoleOutput(...);
$progressWriter = new ConsoleProgressWriter($output, $reader);
// Most useful when added to a workflow
$workflow->addWriter($progressWriter);
There are various optional arguments you can pass to the ConsoleProgressWriter
. These include the output format and
the redraw frequency. You can read more about the options here.
You might want to set the redraw rate higher than the default as it can slow down the import/export process quite a bit
as it will update the console text after every record has been processed by the Workflow
.
$output = new ConsoleOutput(...);
$progressWriter = new ConsoleProgressWriter($output, $reader, 'debug', 100);
Above we set the output format to 'debug' and the redraw rate to 100. This will only re-draw the console progress text after every 100 records.
The debug
format is default as it displays ETA's and Memory Usage. You can use a more simple formatter if you wish:
$output = new ConsoleOutput(...);
$progressWriter = new ConsoleProgressWriter($output, $reader, 'normal', 100);
Instead of implementing your own writer, you can use the quick solution the CallbackWriter offers:
use Ddeboer\DataImport\Writer\CallbackWriter;
$workflow->addWriter(new CallbackWriter(function ($row) use ($storage) {
$storage->store($row);
}));
Instead of implementing your own writer from scratch, you can use AbstractStreamWriter as a basis,
implemented the writeItem
method and you're done:
use Ddeboer\DataImport\Writer\AbstractStreamWriter;
class MyStreamWriter extends AbstractStreamWriter
{
public function writeItem(array $item)
{
fputs($this->getStream(), implode(',', $item));
}
}
$writer = new MyStreamWriter(fopen('php://temp', 'r+'));
$writer->setCloseStreamOnFinish(false);
$workflow->addWriter(new MyStreamWriter());
$workflow->process();
$stream = $writer->getStream();
rewind($stream);
echo stream_get_contents($stream);
Suppose you have 2 stream writers handling fields differently according to one of the fields.
You should then use StreamMergeWriter
to call the appropriate Writer for you.
The default field name is discr
but could be changed with the setDiscriminantField()
method.
use Ddeboer\DataImport\Writer\StreamMergeWriter;
$writer = new StreamMergeWriter();
$writer->addWriter('first writer', new MyStreamWriter());
$writer->addWriter('second writer', new MyStreamWriter());
Build your own writer by implementing the WriterInterface.
A filter decides whether data input is accepted into the import process.
The CallbackFilter wraps your callback function that determines whether
data should be accepted. The data input is accepted only if the function
returns true
.
use Ddeboer\DataImport\Filter\CallbackFilter;
// Don’t import The Beatles
$filter = new CallbackFilter(function ($data) {
return ('The Beatles' != $data['name']);
});
$workflow->addFilter($filter);
OffsetFilter allows you to
- skip a certain amount of items from the beginning
- process only specified amount of items (and skip the rest)
You can combine these two parameters to process a slice from the middle of the data, like rows 5-7 of a CSV file with ten rows.
OffsetFilter is configured by its constructor:
new OffsetFilter($offset = 0, $limit = null)
. Note: $offset
is a 0-based index.
use Ddeboer\DataImport\Filter\OffsetFilter;
// Default implementation is to start from the beginning without maximum count
$filter = new OffsetFilter(0, null);
$filter = new OffsetFilter(); // You can omit both parameters
// Start from the third item, process to the end
$filter = new OffsetFilter(2, null);
$filter = new OffsetFilter(2); // You can omit the second parameter
// Start from the first item, process max three items
$filter = new OffsetFilter(0, 3);
// Start from the third item, process max five items (items 3 - 7)
$filter = new OffsetFilter(2, 5);
This filter is useful if you want to do incremental imports. Specify a threshold
DateTime
instance, a column name (defaults to updated_at
), and a
DateTimeValueConverter
that will be used to convert values read from the
filtered items. The item strictly older than the threshold will be discarded.
use Ddeboer\DataImport\Filter\DateTimeThresholdFilter;
use Ddeboer\DataImport\ValueConverter\DateTimeValueConverter;
new DateTimeThresholdFilter(
new DateTimeValueConverter(),
new \DateTime('yesterday')
);
It’s a common use case to validate the data before you save it to the database. This is exactly what the ValidatorFilter does. To use it, include Symfony’s Validator component in your project:
$ compose require symfony/validator
The ValidatorFilter works as follows:
use Ddeboer\DataImport\Filter\ValidatorFilter;
$filter = new ValidatorFilter($validator);
$filter->add('email', new Assert\Email());
$filter->add('sku', new Assert\NotBlank());
The default behaviour for the validator is to collect all violations and skip
each invalid row. If you want to stop on the first failing row you can call
ValidatorFilter::throwExceptions()
, which throws a
ValidationException
containing the line number and the violation list.
Use the MappingItemConverter to add mappings to your workflow. Your keys from the input data will be renamed according to these mappings. Say you have input data:
$data = array(
array(
'foo' => 'bar',
'baz' => array(
'some' => 'value'
)
)
);
You can map the keys foo
and baz
in the following way:
use Ddeboer\DataImport\ItemConverter\MappingItemConverter;
$converter = new MappingItemConverter();
$converter
->addMapping('foo', 'fooloo')
->addMapping('baz', array('some' => 'else'));
$workflow->addItemConverter($converter)
->process();
Your output data will now be:
array(
array(
'fooloo' => 'bar',
'baz' => array(
'else' => 'value'
)
)
);
Use the NestedMappingItemConverter to add mappings to your workflow if the input data contains nested arrays. Your keys from the input data will be renamed according to these mappings. Say you have input data:
$data = array(
'foo' => 'bar',
'baz' => array(
array(
'another' => 'thing'
),
array(
'another' => 'thing2'
),
)
);
You can map the keys another
in the following way.
use Ddeboer\DataImport\ItemConverter\NestedMappingItemConverter;
$mappings = array(
'foo' => 'foobar',
'baz' => array(
'another' => 'different_thing'
)
);
$converter = new NestedItemMappingConverter('baz');
$converter->addMapping($mappings);
$workflow->addItemConverter($converter)
->process();
Your output data will now be:
array(
'foobar' => 'bar',
'baz' => array(
array(
'different_thing' => 'thing'
),
array(
'different_thing' => 'thing2'
),
)
);
Implement ItemConverterInterface
to create your own item converter:
use Ddeboer\DataImport\ItemConverter\ItemConverterInterface;
class MyItemConverter implements ItemConverterInterface
{
public function convert($item)
{
// Do your conversion and return updated $item
return $changedItem;
}
}
Instead of implementing your own item converter, you can use a callback:
use Ddeboer\DataImport\ItemConverter\CallbackItemConverter;
// Use a fictional $translator service to translate each value
$converter = new CallbackItemConverter(function ($item) use ($translator) {
foreach ($item as $key => $value) {
$item[$key] = $translator->translate($value);
}
return $item;
});
Value converters are used to convert specific fields (e.g., columns in database).
There are two uses for the StringToDateTimeValueConverter:
- Convert a date representation in a format you specify into a
DateTime
object. - Convert a date representation in a format you specify into a different format.
use Ddeboer\DataImport\ValueConverter\StringDateTimeValueConverter;
$converter = new StringToDateTimeValueConverter('d/m/Y H:i:s');
$workflow->addValueConverter('my_date_field', $converter);
If your date string is in a format specified at: http://www.php.net/manual/en/datetime.formats.date.php then you can omit the format parameter.
use Ddeboer\DataImport\ValueConverter\StringToDateTimeValueConverter;
$converter = new StringToDateTimeValueConverter();
$workflow->addValueConverter('my_date_field', $converter);
use Ddeboer\DataImport\ValueConverter\StringToDateTimeValueConverter;
$converter = new StringToDateTimeValueConverter('d/m/Y H:i:s', 'd-M-Y');
$workflow->addValueConverter('my_date_field', $converter);
If your date is in a format specified at: http://www.php.net/manual/en/datetime.formats.date.php you can pass null
as the first argument.
use Ddeboer\DataImport\ValueConverter\StringToDateTimeValueConverter;
$converter = new StringToDateTimeValueConverter(null, 'd-M-Y');
$workflow->addValueConverter('my_date_field', $converter);
The main use of DateTimeToStringValueConverter is to convert DateTime object into it's string representation in proper format. Default format is 'Y-m-d H:i:s';
use Ddeboer\DataImport\ValueConverter\DateTimeToStringValueConverter;
$converter = new DateTimeToStringValueConverter;
$converter->convert(\DateTime('2010-01-01 01:00:00')); //will return string '2010-01-01 01:00:00'
Converts an object into a scalar value. To use this converter, you must include Symfony’s PropertyAccess component in your project:
$ composer require symfony/property-access
If your object has a __toString()
method, that value will be used:
use Ddeboer\DataImport\ValueConverter\ObjectConverter;
class SecretAgent
{
public function __toString()
{
return '007';
}
}
$converter = new ObjectConverter();
$string = $converter->convert(new SecretAgent()); // $string will be '007'
If your object has no __toString()
method, its accessors will be called
instead:
class Villain
{
public function getName()
{
return 'Bad Guy';
}
}
class Organization
{
public function getVillain()
{
return new Villain();
}
}
use Ddeboer\DataImport\ValueConverter\ObjectConverter;
$converter = new ObjectConverter('villain.name');
$string = $converter->convert(new Organization()); // $string will be 'Bad Guy'
Looks up an object in the database based on a string value:
use Ddeboer\DataImport\ValueConverter\StringToObjectConverter;
$converter = new StringToObjectConverter($repository, 'name');
$workflow->addValueConverter('input_name', $converter);
Use this if you want to save the trouble of writing a dedicating class:
use Ddeboer\DataImport\ValueConverter\CallbackValueConverter;
$callable = function ($item) {
return implode(',', $item);
};
$converter = new CallbackValueConverter($callable);
$output = $converter->convert(array('foo', 'bar')); // $output will be "foo,bar"
Looks for a key in a hash you must provide in the constructor:
use Ddeboer\DataImport\ValueConverter\MappingValueConverter;
$converter = new MappingValueConverter(array(
'source' => 'destination'
));
$converter->convert('source'); // destination
$converter->convert('unexpected value'); // throws an UnexpectedValueException
This example shows how you can read data from a CSV file and write that to the database.
Assume we have the following CSV file:
event;beginDate;endDate
Christmas;20131225;20131226
New Year;20131231;20140101
And we want to write this data to a Doctrine entity:
namespace MyApp;
use Doctrine\ORM\Mapping as ORM;
/**
* @ORM\Entity
*/
class Event
{
/**
* @ORM\Column()
*/
protected $event;
/**
* @ORM\Column(type="datetime")
*/
protected $beginDate;
/**
* @ORM\Column(type="datetime")
*/
protected $endDate;
public function setEvent($event)
{
$this->event = $event;
}
public function setBeginDate($date)
{
$this->beginDate = $date;
}
public function setEndDate($date)
{
$this->endDate = $date;
}
// And some getters
}
Then you can import the CSV and save it as your entity in the following way.
use Ddeboer\DataImport\Workflow;
use Ddeboer\DataImport\Reader\CsvReader;
use Ddeboer\DataImport\Writer\DoctrineWriter;
use Ddeboer\DataImport\ValueConverter\StringToDateTimeValueConverter;
// Create and configure the reader
$file = new \SplFileObject('input.csv');
$csvReader = new CsvReader($file);
// Tell the reader that the first row in the CSV file contains column headers
$csvReader->setHeaderRowNumber(0);
// Create the workflow from the reader
$workflow = new Workflow($csvReader);
// Create a writer: you need Doctrine’s EntityManager.
$doctrineWriter = new DoctrineWriter($entityManager, 'MyApp:Event');
$workflow->addWriter($doctrineWriter);
// Add a converter to the workflow that will convert `beginDate` and `endDate`
// to \DateTime objects
$dateTimeConverter = new StringToDateTimeValueConverter('Ymd');
$workflow
->addValueConverter('beginDate', $dateTimeConverter)
->addValueConverter('endDate', $dateTimeConverter);
// Process the workflow
$workflow->process();
This example shows how you can export data to a CSV file.
use Ddeboer\DataImport\Workflow;
use Ddeboer\DataImport\Reader\ArrayReader;
use Ddeboer\DataImport\Writer\CsvWriter;
use Ddeboer\DataImport\ValueConverter\CallbackValueConverter;
// Your input data
$reader = new ArrayReader(array(
array(
'first', // This is for the CSV header
'last',
array(
'first' => 'james',
'last' => 'Bond'
),
array(
'first' => 'hugo',
'last' => 'Drax'
)
))
);
// Create the workflow from the reader
$workflow = new Workflow($reader);
// Add the writer to the workflow
$file = new \SplFileObject('output.csv', 'w');
$writer = new CsvWriter($file);
$workflow->addWriter($writer);
// As you can see, the first names are not capitalized correctly. Let's fix
// that with a value converter:
$converter = new CallbackValueConverter(function ($input) {
return ucfirst($input);
});
$workflow->addValueConverter('first', $converter);
// Process the workflow
$workflow->process();
This will write a CSV file output.csv
where the first names are capitalized:
first;last
James;Bond
Hugo;Drax
The ArrayValueConverterMap is used to filter values of a multi-level array.
The converters defined in the list are applied on every data-item's value that match the defined array_keys.
//...
$data = array(
'products' => array(
0 => array(
'name' => 'some name',
'price' => '€12,16',
),
1 => array(
'name' => 'some name',
'price' => '€12,16',
)
)
);
// ...
// create the workflow and reader etc.
// ...
$workflow->addValueConverter(new ArrayValueConverterMap(array(
'name' => array(new CharsetValueConverter('UTF-8', 'UTF-16')), // encode to UTF-8
'price' => array(new CallbackValueConverter(function ($input) {
return str_replace('€', '', $input); // remove € char
}),
)));
// ..
// after filtering data looks as follows
$data = array(
'products' => array(
0 => array(
'name' => 'some name', // in UTF-8
'price' => '12,16',
),
1 => array(
'name' => 'some name',
'price' => '12,16',
)
)
);
Clone this repository:
$ git clone https://github.com/ddeboer/data-import.git
$ cd data-import
Install dev dependencies:
$ composer install --dev
And run PHPUnit:
$ phpunit
DataImport is released under the MIT license. See the LICENSE file for details.