conflict_resolution stringlengths 27 16k |
|---|
<<<<<<<
if (type == null)
throw new BadParameterEx("type", null);
=======
if (type == null) {
throw new BadParameterEx("type", type);
}
>>>>>>>
if (type == null) {
throw new BadParameterEx("type", null);
}
<<<<<<<
//-------------------------------------------------------------------... |
<<<<<<<
import org.fao.geonet.services.api.exception.NoResultsFoundException;
=======
import org.fao.geonet.exceptions.UserNotFoundEx;
>>>>>>>
import org.fao.geonet.exceptions.UserNotFoundEx;
import org.fao.geonet.services.api.exception.NoResultsFoundException; |
<<<<<<<
import jeeves.server.context.ServiceContext;
=======
import java.net.MalformedURLException;
import java.net.URL;
import java.nio.file.Path;
import java.util.Collections;
import java.util.HashSet;
import java.util.LinkedList;
import java.util.List;
import java.util.Map;
import java.util.Set;
import java.util.... |
<<<<<<<
import static com.google.common.io.Files.getNameWithoutExtension;
import static org.fao.geonet.api.ApiParams.API_PARAM_RECORD_UUID;
import static org.fao.geonet.api.records.formatters.FormatterConstants.SCHEMA_PLUGIN_FORMATTER_DIR;
import static org.springframework.data.jpa.domain.Specifications.where;
import ... |
<<<<<<<
import org.fao.geonet.Logger;
import org.fao.geonet.domain.Source;
import org.fao.geonet.exceptions.BadInputEx;
=======
import jeeves.server.resources.ResourceManager;
import org.fao.geonet.constants.Geonet;
>>>>>>>
<<<<<<<
import org.fao.geonet.repository.SourceRepository;
=======
import org.fao.geonet.... |
<<<<<<<
import java.io.IOException;
import java.io.InputStream;
import java.io.OutputStream;
import java.nio.file.DirectoryStream;
import java.nio.file.Files;
import java.nio.file.Path;
import java.nio.file.attribute.FileTime;
import java.util.ArrayList;
import java.util.HashMap;
import java.util.HashSet;
import java.u... |
<<<<<<<
*
*
* @param context
* @throws IOException
* @throws JDOMException
=======
>>>>>>>
<<<<<<<
host = sm.getValue(Settings.SYSTEM_FEEDBACK_MAILSERVER_HOST);
port = sm.getValue(Settings.SYSTEM_FEEDBACK_MAILSERVER_PORT);
from = sm.getValue(Settings.SYSTEM_FEEDBACK_... |
<<<<<<<
public class XmlSearch implements Service {
private ServiceConfig _config;
private String _searchFast; //true, false, index
//--------------------------------------------------------------------------
//---
//--- Init
//---
//---------------------------------------------------------... |
<<<<<<<
=======
public static final String SYSTEM_SITE_SITE_ID_PATH = "system/site/siteId";
public static final String SYSTEM_SITE_NAME_PATH = "system/site/name";
public static final String SYSTEM_SITE_LABEL_PREFIX = "system/site/labels/";
public static final String CSW_TRANSACTION_XPATH_UPDATE_CREATE... |
<<<<<<<
public ObjectReader at(String value) {
return new ObjectReader(this, new JsonPointerBasedFilter(value));
=======
public ObjectReader at(final String pointerExpr) {
_assertNotNull("pointerExpr", pointerExpr);
return new ObjectReader(this, new JsonPointerBasedFilter(pointerExpr))... |
<<<<<<<
@EntityListeners(GroupEntityListenerManager.class)
=======
@SequenceGenerator(name=Group.ID_SEQ_NAME, initialValue=100, allocationSize=1)
>>>>>>>
@EntityListeners(GroupEntityListenerManager.class)
@SequenceGenerator(name=Group.ID_SEQ_NAME, initialValue=100, allocationSize=1)
<<<<<<<
=======
static fin... |
<<<<<<<
for (OperationAllowed opAllowed : opsAllowed) {
if (opAllowed.getId().getOperationId() != ReservedOperation.notify.getId())
continue;
=======
for (OperationAllowed opAllowed : opsAllowed) {
if (opAllowed.getId().getOperationId() != ReservedOperation.notify.... |
<<<<<<<
import com.fasterxml.jackson.annotation.JsonIgnore;
import com.fasterxml.jackson.annotation.JsonProperty;
import com.fasterxml.jackson.annotation.JsonPropertyOrder;
=======
>>>>>>>
import com.fasterxml.jackson.annotation.JsonIgnore;
import com.fasterxml.jackson.annotation.JsonProperty;
import com.fasterx... |
<<<<<<<
private static final Random RANDOM = new Random();
public static String randomId() {
return "N" + RANDOM.nextInt(Integer.MAX_VALUE);
}
=======
public static String getMax(Object values) {
String[] strings = values.toString().split(" ");
String max = "";
... |
<<<<<<<
@Component(CatalogService.BEAN_PREFIX+GetCapabilities.NAME)
public class GetCapabilities extends AbstractOperation implements CatalogService
{
//---------------------------------------------------------------------------
//---
//--- Constructor
//---
//---------------------------------------------------... |
<<<<<<<
@EntityListeners(CswCapabilitiesInfoFieldEntityListenerManager.class)
=======
@SequenceGenerator(name=CswCapabilitiesInfoField.ID_SEQ_NAME, initialValue=100, allocationSize=1)
>>>>>>>
@EntityListeners(CswCapabilitiesInfoFieldEntityListenerManager.class)
@SequenceGenerator(name=CswCapabilitiesInfoField.ID_SEQ... |
<<<<<<<
import org.fao.geonet.kernel.setting.SettingManager;
import org.fao.geonet.utils.Xml;
import org.fao.geonet.exceptions.BadParameterEx;
import org.fao.geonet.exceptions.OperationAbortedEx;
=======
>>>>>>>
<<<<<<<
import org.fao.geonet.utils.*;
=======
>>>>>>>
<<<<<<<
import org.fao.geonet.languages.Iso... |
<<<<<<<
import org.fao.geonet.repository.SourceRepository;
=======
import org.fao.geonet.kernel.harvest.harvester.HarvestResult;
import org.fao.geonet.lib.Lib;
>>>>>>>
import org.fao.geonet.repository.SourceRepository;
<<<<<<<
Harvester h = new Harvester(log, context, params);
result = h.harvest();
=======
... |
<<<<<<<
@EntityListeners(value = {UserEntityListenerManager.class})
=======
@SequenceGenerator(name=User.ID_SEQ_NAME, initialValue=100, allocationSize=1)
>>>>>>>
@EntityListeners(value = {UserEntityListenerManager.class})
@SequenceGenerator(name=User.ID_SEQ_NAME, initialValue=100, allocationSize=1) |
<<<<<<<
settingMan.add("id:"+siteId, "capabUrl", params.capabUrl);
settingMan.add("id:"+siteId, "icon", params.icon);
settingMan.add("id:"+siteId, "rejectDuplicateResource", params.rejectDuplicateResource);
=======
settingMan.add(dbms, "id:"+siteId, "capabUrl", params.capabUrl);
settingMan... |
<<<<<<<
import com.google.common.annotations.VisibleForTesting;
=======
>>>>>>>
import com.google.common.annotations.VisibleForTesting;
<<<<<<<
=======
import java.io.File;
>>>>>>> |
<<<<<<<
try{
=======
try {
>>>>>>>
try {
<<<<<<<
boolean localRating = settingManager.getValueAsBool(Settings.SYSTEM_LOCALRATING_ENABLE, false);
=======
boolean localRating = settingManager.getValueAsBool("system/localrating/enable", ... |
<<<<<<<
List<Pair<String, Pair<Document, List<CategoryPath>>>> docs = buildIndexDocument(schemaDir, metadata, id, moreFields, metadataType, false);
=======
List<Pair<String, Pair<Document, Collection<CategoryPath>>>> docs = buildIndexDocument(schemaDir, metadata, id, moreFields, metadataType, title, f... |
<<<<<<<
import jeeves.server.sources.http.ServletPathFinder;
import org.fao.geonet.NodeInfo;
=======
>>>>>>>
<<<<<<<
import java.sql.SQLException;
import java.util.HashMap;
import java.util.List;
import java.util.Map;
import java.util.NoSuchElementException;
import javax.annotation.Nonnull;
import javax.annotation.... |
<<<<<<<
=======
import org.apache.log4j.Priority;
import org.fao.geonet.GeonetContext;
>>>>>>>
import org.apache.log4j.Priority;
<<<<<<<
public Element removeHiddenElements(boolean isIndexingTask, Metadata metadata) throws Exception {
String id = String.valueOf(metadata.getId());
Element metadat... |
<<<<<<<
public void doHarvest(Logger log) throws Exception {
Harvester h = new Harvester(log, context, params);
serverResults = h.harvest();
=======
protected void doHarvest(Logger log, ResourceManager rm) throws Exception {
Dbms dbms = (Dbms) rm.open(Geonet.Res.MAIN_DB);
h = new Harvester(log, context, db... |
<<<<<<<
=======
// TODO add group to user
//String group = getHeader(req, config.getGroupKey(), "");
>>>>>>>
// TODO add group to user
//String group = getHeader(req, config.getGroupKey(), "");
<<<<<<<
=======
// TODO add group to user
//if (group.equals("")) {
// group = config.getDefaultGroup(... |
<<<<<<<
=======
import jeeves.utils.Util;
import jeeves.utils.Xml;
import org.eclipse.emf.common.command.AbortExecutionException;
>>>>>>>
import org.eclipse.emf.common.command.AbortExecutionException;
<<<<<<<
public HarvestResult harvest() throws Exception {
ListIdentifiersRequest req = new ListIdentifiersReq... |
<<<<<<<
import java.io.IOException;
import java.io.InputStream;
import java.io.OutputStream;
import java.nio.file.DirectoryStream;
import java.nio.file.Files;
import java.nio.file.Path;
import java.nio.file.attribute.FileTime;
import java.util.ArrayList;
import java.util.HashMap;
import java.util.HashSet;
import java.u... |
<<<<<<<
SettingManager settingMan = context.getApplicationContext().getBean(SettingManager.class);
// --- Migrate database if an old one is found
migrateDatabase(servletContext, dbms, settingMan, version, subVersion, context.getAppPath());
=======
SettingManager settingMan = null;
HarvesterSettingsManager... |
<<<<<<<
import org.fao.geonet.repository.AbstractSpringDataTest;
import org.fao.geonet.repository.SourceRepository;
import org.fao.geonet.repository.UserRepository;
=======
import org.fao.geonet.kernel.search.LuceneConfig;
import org.fao.geonet.kernel.search.SearchManager;
import org.fao.geonet.kernel.search.index.Di... |
<<<<<<<
import org.fao.geonet.util.PasswordUtil;
=======
import org.fao.geonet.util.MailUtil;
>>>>>>>
import org.fao.geonet.util.PasswordUtil;
import org.fao.geonet.util.MailUtil;
<<<<<<<
if (!sendRegistrationEmail(params, password, host, port, from, thisSite, siteURL)) {
return element.addContent(new E... |
<<<<<<<
* When a remote rating is applied, the local rating is not updated. It will be updated
* on the next harvest run (FIXME ?).
=======
* When a remote rating is applied, the local rating is not updated. It will be updated on the next
* harvest run (FIXME ?).
>>>>>>>
* When a remote rating is applied, the l... |
<<<<<<<
import org.fao.geonet.NodeInfo;
import org.fao.geonet.kernel.search.index.IndexingList;
import org.fao.geonet.kernel.search.index.IndexingTask;
import org.fao.geonet.repository.specification.*;
import org.fao.geonet.repository.statistic.PathSpec;
import org.fao.geonet.util.FileCopyMgr;
import org.fao.geonet.ut... |
<<<<<<<
import org.fao.geonet.Logger;
=======
>>>>>>>
import org.fao.geonet.Logger;
<<<<<<<
protected void addCategories(String id, Iterable<String> categories, CategoryMapper localCateg, DataManager dataMan, ServiceContext context, Logger log, String serverCategory) throws Exception {
=======
public vo... |
<<<<<<<
Format format, boolean skipUUID, boolean resolveXlink, boolean removeXlinkAttribute) throws Exception {
Pair<AbstractMetadata, String> recordAndMetadata =
MEFLib.retrieveMetadata(context, uuid, resolveXlink, removeXlinkAttribute);
AbstractMetadata reco... |
<<<<<<<
=======
import static org.quartz.JobKey.jobKey;
import java.io.File;
import java.lang.reflect.Method;
import java.sql.SQLException;
import java.text.SimpleDateFormat;
import java.util.Date;
import java.util.HashMap;
import java.util.LinkedList;
import java.util.List;
import java.util.Map;
import java.util.Se... |
<<<<<<<
import org.fao.geonet.repository.SourceRepository;
=======
import org.fao.geonet.kernel.harvest.harvester.HarvestResult;
import org.fao.geonet.lib.Lib;
>>>>>>>
import org.fao.geonet.repository.SourceRepository;
<<<<<<<
Harvester h = new Harvester(log, context, params);
result = h.harvest();
=======
... |
<<<<<<<
=======
import com.vividsolutions.jts.util.Assert;
import org.apache.lucene.document.Document;
import org.fao.geonet.entitylistener.MetadataEntityListenerManager;
import org.fao.geonet.utils.Xml;
import org.hibernate.annotations.Type;
import org.jdom.Element;
import org.jdom.JDOMException;
import org.jdom.ou... |
<<<<<<<
this.metadataUtils = context.getBean(IMetadataUtils.class);
this.settingMan = context.getBean(HarvesterSettingsManager.class);
=======
this.harvesterSettingsManager = context.getBean(HarvesterSettingsManager.class);
this.settingManager = context.getBean(SettingManager.class);
... |
<<<<<<<
MetadataType.METADATA);
=======
MetadataType.METADATA, metadata.getDataInfo().getTitle(), false);
>>>>>>>
MetadataType.METADATA, false); |
<<<<<<<
final ServiceContext context = createServiceContext(locale.getISO3Language(),
formatType, request.getNativeRequest(HttpServletRequest.class));
AbstractMetadata metadata = ApiUtils.canViewRecord(metadataUuid, servletRequest);
=======
final String language = LanguageUtils.lo... |
<<<<<<<
import org.fao.geonet.domain.AbstractMetadata;
=======
import org.fao.geonet.domain.Metadata;
import org.fao.geonet.domain.userfeedback.RatingsSetting;
>>>>>>>
import org.fao.geonet.domain.AbstractMetadata;
import org.fao.geonet.domain.userfeedback.RatingsSetting; |
<<<<<<<
@EntityListeners(MetadataEntityListenerManager.class)
=======
@SequenceGenerator(name=Metadata.ID_SEQ_NAME, initialValue=100, allocationSize=1)
>>>>>>>
@EntityListeners(MetadataEntityListenerManager.class)
@SequenceGenerator(name=Metadata.ID_SEQ_NAME, initialValue=100, allocationSize=1)
<<<<<<<
=======
... |
<<<<<<<
import org.fao.geonet.constants.Geonet;
import org.fao.geonet.utils.BinaryFile;
=======
import org.fao.geonet.constants.Geonet;
>>>>>>>
import org.fao.geonet.constants.Geonet;
<<<<<<<
import java.io.File;
import java.io.IOException;
=======
import java.io.IOException;
import java.nio.file.DirectoryStrea... |
<<<<<<<
//---------------------------------------------------------------------------
//---
//--- API methods
//---
//---------------------------------------------------------------------------
public String getSiteName()
{
=======
//---------------------------------------------------------------------... |
<<<<<<<
ObjectMapper mapper = jsonMapperBuilder()
.enableDefaultTypingAsProperty(NoCheckSubTypeValidator.instance,
DefaultTyping.NON_FINAL, "@class")
=======
ObjectMapper om = JsonMapper.builder()
.activateDefaultTypingAsProperty(NoCheckSubTypeVa... |
<<<<<<<
=======
import jeeves.utils.Xml;
>>>>>>>
<<<<<<<
import java.util.*;
=======
import java.util.HashMap;
import java.util.HashSet;
import java.util.LinkedList;
import java.util.List;
import java.util.Map;
import java.util.Set;
>>>>>>>
import java.util.*;
<<<<<<<
if(log.isDebugEnabled()) log.... |
<<<<<<<
import java.io.IOException;
import java.io.InputStream;
import java.io.OutputStream;
import java.net.URL;
import java.nio.file.Files;
import java.nio.file.Path;
import java.util.ArrayList;
import java.util.Collections;
import java.util.HashMap;
import java.util.Iterator;
import java.util.LinkedList;
import java... |
<<<<<<<
final List<Object> nodeList = trySelectNode(metadataRecord, metadataSchema, xpathProperty, true).results;
=======
// Removes root metadata element for xpath filters
xpathProperty = cleanRootFromXPath(xpathProperty, metadataRecord);
final Object propNode = trySe... |
<<<<<<<
public static SystemInfo createForTesting(String stagingProfile) {
return new SystemInfo(stagingProfile, "testing", "3.0.0", "SNAPSHOT", "testing", "testing", "testing", "", "");
}
=======
>>>>>>>
public static SystemInfo createForTesting(String stagingProfile) {
return new Syste... |
<<<<<<<
JsonPOJOBuilder.Value builderConfig = (ai == null) ? null : ai.findPOJOBuilderConfig(config, builderClass);
String mutatorPrefix = (builderConfig == null) ? JsonPOJOBuilder.DEFAULT_WITH_PREFIX : builderConfig.withPrefix;
=======
JsonPOJOBuilder.Value builderConfig = (ai == ... |
<<<<<<<
Object filterId = findFilterId(config, beanDesc);
MapSerializer mapSer = MapSerializer.construct(config.getAnnotationIntrospector().findPropertiesToIgnore(beanDesc.getClassInfo()),
=======
*/
Object filterId = findFilterId(config, beanDesc);
... |
<<<<<<<
=======
// 24-May-2012, tatu: Comment out for 2.0.x to keep tests green; leave for 2.1 to fix
/*
// [JACKSON-822]: ensure that type can be coerced
public void testTypedLists() throws Exception
{
ObjectMapper mapper = new ObjectMapper();
List<Issue822Interface> list = ... |
<<<<<<<
=======
import java.util.ArrayList;
import java.util.Collection;
>>>>>>>
import java.util.ArrayList;
import java.util.Collection;
<<<<<<<
import org.apache.zeppelin.user.AuthenticationInfo;
=======
import org.apache.zeppelin.scheduler.JobListener;
import org.apache.zeppelin.search.SearchService;
>>>>>>>
... |
<<<<<<<
import se.kth.meta.entity.Template;
import se.kth.meta.exception.DatabaseException;
=======
import se.kth.hopsworks.controller.FolderNameValidator;
>>>>>>>
<<<<<<<
=======
private boolean createDataset(String dsPath, Inode parent, String dsName,
int template) throws AppException {
boolean... |
<<<<<<<
@NotNull
@Size(min = 1, max = 30)
@Column(name = "ethical_satus")
private String ethicalStatus;
@OneToMany(cascade = CascadeType.ALL, mappedBy = "trackStudy")
private Collection<StudyGroups> studyGroupsCollection;
=======
>>>>>>>
@NotNull
@Size(min = 1, max = 30)
@Colu... |
<<<<<<<
private String restEndpoint;
=======
private String jobName;
private ElasticProperties elastic;
>>>>>>>
private String restEndpoint;
private String jobName;
private ElasticProperties elastic;
<<<<<<<
Integer projectId, String projectName, String restEndPoint) {
=======
In... |
<<<<<<<
// spark 1.5.x replaced --num-executors with --properties-file
// amargs.append(" --num-executors ").append(numberOfExecutors);
=======
// https://fossies.org/diffs/spark/1.4.1_vs_1.5.0/
// yarn/src/main/scala/org/apache/spark/deploy/yarn/
// ApplicationMasterArguments.scala-diff.html
... |
<<<<<<<
private static final String VARIABLE_ANACONDA_DIR = "anaconda_dir";
=======
private static final String VARIABLE_INFLUXDB_ADDRESS = "influxdb_address";
private static final String VARIABLE_INFLUXDB_USER = "influxdb_user";
private static final String VARIABLE_INFLUXDB_PW = "influxdb_pw";
>>>>>>>
... |
<<<<<<<
import javax.servlet.ServletContext;
=======
import javax.inject.Inject;
>>>>>>>
import javax.servlet.ServletContext;
import javax.inject.Inject;
<<<<<<<
import se.kth.meta.db.Dbao;
import se.kth.meta.entity.Templates;
import se.kth.meta.exception.DatabaseException;
=======
import se.kth.hopsworks.control... |
<<<<<<<
builder.addLocalResource(dto, !appPath.startsWith("hdfs:"));
builder.addToAppMasterEnvironment(YarnRunner.KEY_CLASSPATH,
dto.getName());
extraClassPathFiles.append(dto.getName()).append(File.pathSeparator);
=======
if (dto.getName().equals(Settings.K_CERTIFICATE) || dto.getNa... |
<<<<<<<
ObjectMapper mapper = jsonMapperBuilder()
.enableDefaultTyping(NoCheckSubTypeValidator.instance,
DefaultTyping.NON_FINAL, JsonTypeInfo.As.PROPERTY)
=======
final ObjectMapper mapper = jsonMapperBuilder()
.activateDefaultTyping(NoCheckSubT... |
<<<<<<<
resources.add(se.kth.hopsworks.rest.DataSetService.class);
resources.add(se.kth.hopsworks.rest.ProjectMembers.class);
=======
resources.add(se.kth.hopsworks.rest.CuneiformService.class);
resources.add(se.kth.hopsworks.rest.DataSetService.class);
resources.add(se.kth.hopsworks.rest.JobServi... |
<<<<<<<
@NamedQuery(name = "Consent.findAll",
query = "SELECT c FROM Consent c"),
@NamedQuery(name = "Consent.findById",
query = "SELECT c FROM Consent c WHERE c.id = :id"),
@NamedQuery(name = "Consent.findByDate",
query = "SELECT c FROM Consent c WHERE c.date = :date"),
@NamedQuer... |
<<<<<<<
SPARK_HISTORY_SERVER_IP = setIpVar(VARIABLE_SPARK_HISTORY_SERVER_IP, SPARK_HISTORY_SERVER_IP);
ZK_IP = setIpVar(VARIABLE_ZK_IP, ZK_IP);
ZK_USER = setUserVar(VARIABLE_ZK_USER, ZK_USER);
ZK_DIR = setDirVar(VARIABLE_ZK_DIR, ZK_DIR);
KAFKA_USER = setUserVar(VARIABLE_KAFKA_USER, KAFKA_... |
<<<<<<<
fileOps.copyToHDFSFromLocal(true, new File(stagingManager.getStagingPath(), info.getResumableFilename()).
getAbsolutePath(), uploadPath
+ info.getResumableFilename());
=======
fileOps.copyToHDFSFromLocal(true, new File(stagingManager.
getStagingPath(... |
<<<<<<<
=======
} catch (IOException e) {
throw new AppException(Response.Status.INTERNAL_SERVER_ERROR.
getStatusCode(), e.getMessage());
>>>>>>> |
<<<<<<<
private String name;
private boolean dir;
private boolean parent;
private String path;
private Date modification;
private int id;
private int template;
public InodeView() {
}
=======
private String name;
private boolean dir;
private boolean parent;
private String path;
privat... |
<<<<<<<
=======
import javax.ejb.TransactionAttribute;
import javax.ejb.TransactionAttributeType;
import se.kth.bbc.jobs.jobhistory.ExecutionInputfilesFacade;
>>>>>>>
import javax.ejb.TransactionAttribute;
import javax.ejb.TransactionAttributeType; |
<<<<<<<
import java.io.BufferedReader;
import java.io.FileNotFoundException;
import java.io.InputStreamReader;
import java.io.OutputStreamWriter;
import java.net.HttpURLConnection;
import java.net.MalformedURLException;
import java.net.URL;
=======
import java.util.Collections;
import java.util.Comparator;
>>>>>>>
i... |
<<<<<<<
import se.kth.hopsworks.hdfs.fileoperations.DistributedFsService;
=======
import se.kth.hopsworks.hdfsUsers.controller.HdfsUsersController;
>>>>>>>
import se.kth.hopsworks.hdfs.fileoperations.DistributedFsService;
import se.kth.hopsworks.hdfsUsers.controller.HdfsUsersController;
<<<<<<<
@EJB
private Dis... |
<<<<<<<
=======
public String getOwnerRole() {
return owner;
}
public void setOwnerRole(String owner) {
this.owner = owner;
}
public String getNewTeamRole() {
return newTeamRole;
}
public void setNewTeamRole(String newTeamRole) {
this.ne... |
<<<<<<<
private boolean enableLogDir = true;
private String eventLogDir;
=======
private String sessionId;//used by Kafka
>>>>>>>
private boolean enableLogDir = true;
private String eventLogDir;
private String sessionId;//used by Kafka
<<<<<<<
// if (classPath == null || classPath.isEmpty()) {
// ... |
<<<<<<<
PaasCredentials credentials = credentialsEJB.find();
Provider check = Provider.fromString(credentials.getProvider());
if (Provider.AWS_EC2.equals(check)) {
provider = Provider.AWS_EC2.toString();
=======
PaaSCredentials credentials = credentialsEJB.find();
P... |
<<<<<<<
=======
private String kafkaAddress;
>>>>>>>
private String kafkaAddress; |
<<<<<<<
import se.kth.bbc.fileoperations.FileSystemOperations;
import se.kth.bbc.lims.Constants;
=======
>>>>>>>
import se.kth.bbc.fileoperations.FileSystemOperations; |
<<<<<<<
private boolean createDataset(String dsPath, Inode parent, String dsName,
int template) throws AppException {
boolean success = false;
try {
success = fileOps.mkDir(dsPath);
//the inode has been created in the file system
if (success && template != 0) {
//get the... |
<<<<<<<
= "SELECT d FROM Dataset d WHERE d.name = :name AND d.project = :projectId")})
=======
= "SELECT d FROM Dataset d WHERE d.name = :name AND d.projectId = :projectId"),
@NamedQuery(name = "Dataset.findSharedWithProject",
query
= "SELECT d FROM Dataset d WHERE d.projectI... |
<<<<<<<
ObjectMapper mapper = jsonMapperBuilder()
.enableDefaultTyping(NoCheckSubTypeValidator.instance)
=======
ObjectMapper m = JsonMapper.builder()
.activateDefaultTyping(NoCheckSubTypeValidator.instance)
>>>>>>>
ObjectMapper mapper = jsonMapperBuilder()
... |
<<<<<<<
_mixIns, rootNames, _configOverrides,
_coercionConfigs);
=======
_subtypeResolver, _mixIns, rootNames, _configOverrides);
>>>>>>>
_subtypeResolver, _mixIns, rootNames, _configOverrides,
_coercionConfigs); |
<<<<<<<
private TornadoCoreRuntime() {
=======
private final OptionValues options;
// @formatter:off
public enum TORNADO_DRIVERS_DESCRIPTION {
OPENCL("implemented"),
PTX("unsupported");
String status;
TORNADO_DRIVERS_DESCRIPTION(String status) {
t... |
<<<<<<<
return ctxt.handleUnexpectedToken(getValueType(ctxt), p.currentToken(), p,
"Cannot deserialize a POJO (of type %s) from non-Array representation (token: %s): "
+"type/property designed to be serialized as JSON Array",
_beanType.getRawClass().getName(),
... |
<<<<<<<
private boolean isOpenCLPreLoadBinary(OCLDeviceContextInterface deviceContext, String deviceInfo) {
=======
@Override
public DeviceBuffer createBuffer(int[] arr) {
return getDeviceContext().getMemoryManager().createDeviceBuffer(arr);
}
private boolean isOpenCLPreLoadBinary(OCLDevi... |
<<<<<<<
boolean isCached(String methodName, SchedulableTask task);
=======
int getDeviceIndex();
String getDeviceName();
>>>>>>>
boolean isCached(String methodName, SchedulableTask task);
int getDeviceIndex();
String getDeviceName(); |
<<<<<<<
import org.graalvm.compiler.lir.Variable;
=======
import uk.ac.manchester.tornado.drivers.opencl.graal.OCLArchitecture;
>>>>>>>
import org.graalvm.compiler.lir.Variable;
import uk.ac.manchester.tornado.drivers.opencl.graal.OCLArchitecture; |
<<<<<<<
public void enableThreadSharing() {
TornadoInternalError.unimplemented();
}
@Override
=======
public int[] checkAtomicsForTask(SchedulableTask task) {
return null;
}
@Override
>>>>>>>
public int[] checkAtomicsForTask(SchedulableTask task) {
return null;
... |
<<<<<<<
long copyInValue = timeProfiler.getTimer(ProfilerType.COPY_IN_TIME);
copyInValue += event.getExecutionTime();
timeProfiler.setTimer(ProfilerType.COPY_IN_TIME, copyInValue);
long dispatchValue = timeProfiler.getTimer... |
<<<<<<<
import jdk.vm.ci.hotspot.HotSpotCallingConventionType;
import jdk.vm.ci.hotspot.HotSpotResolvedJavaField;
import jdk.vm.ci.meta.*;
=======
import static jdk.vm.ci.hotspot.HotSpotJVMCIRuntimeProvider.getArrayBaseOffset;
import static org.graalvm.compiler.nodes.NamedLocationIdentity.ARRAY_LENGTH_LOCATION;
impor... |
<<<<<<<
import uk.ac.manchester.tornado.drivers.common.graal.compiler.DumpLowTierGraph;
=======
import uk.ac.manchester.tornado.api.TornadoDeviceContext;
>>>>>>>
import uk.ac.manchester.tornado.api.TornadoDeviceContext;
import uk.ac.manchester.tornado.drivers.common.graal.compiler.DumpLowTierGraph; |
<<<<<<<
private TornadoCoreRuntime() {
=======
private final OptionValues options;
// @formatter:off
public enum TORNADO_DRIVERS_DESCRIPTION {
OPENCL("implemented"),
PTX("unsupported");
String status;
TORNADO_DRIVERS_DESCRIPTION(String status) {
t... |
<<<<<<<
final long endSequentialCode = (TIME_IN_NS) ? System.nanoTime() : System.currentTimeMillis();
=======
final long endSequentialCode = (TIME_IN_NS) ? System.nanoTime() : System.currentTimeMillis();
>>>>>>>
final long endSequentialCode = (TIME_IN_NS) ? System.nanoTime() : Sy... |
<<<<<<<
import uk.ac.manchester.tornado.api.exceptions.TornadoRuntimeException;
import uk.ac.manchester.tornado.runtime.graal.nodes.TornadoReduceAddNode;
import uk.ac.manchester.tornado.runtime.graal.nodes.TornadoReduceMulNode;
import uk.ac.manchester.tornado.runtime.graal.nodes.TornadoReduceSubNode;
=======
import u... |
<<<<<<<
void useDefaultThreadScheduler(boolean use);
=======
void updateReference(Object oldRef, Object newRef);
>>>>>>>
void updateReference(Object oldRef, Object newRef);
void useDefaultThreadScheduler(boolean use); |
<<<<<<<
private final long id;
private final List<OCLTargetDevice> devices;
=======
private final long contextID;
private final List<OCLDevice> devices;
>>>>>>>
private final long contextID;
private final List<OCLTargetDevice> devices;
<<<<<<<
private static final int MAX_ALLOCATED_REGIO... |
<<<<<<<
System.out.printf("waitlist:\n");
=======
System.out.printf("waitlist:\n");
>>>>>>>
System.out.printf("waitlist:\n");
<<<<<<<
System.out.printf("[%d] 0x%x - %s 0x%x\n",index,events[value],EVENT_DESCRIPTIONS[descriptors[value]], tags[value]);
=======
System.o... |
<<<<<<<
final MultiplexFunctionParameterEditor functionParameterEditor = panel.getFunctionParameterEditor();
final DefaultFunctionParameterEditor activeEditor = functionParameterEditor.getDefaultEditor();
=======
MultiplexFunctionParameterEditor functionParameterEditor = panel.getFunctionParameterEditor()... |
<<<<<<<
=======
>>>>>>>
<<<<<<<
=======
>>>>>>> |
<<<<<<<
public static File createTestOutputFile()
{
return createTestOutputFile(null);
}
public static File createTestOutputFile(String name)
{
final File file = new File("test-output");
//noinspection ResultOfMethodCallIgnored
file.mkdir();
if (StringUtils.isEmpty(name, true))
{
... |
<<<<<<<
public And<StringSubject> contains(String string) {
if (!getSubject().contains(string)) {
=======
public StringSubject contains(String string) {
if (getSubject() == null) {
if (string != null) {
fail("contains", string);
}
} else if (!getSubject().contains(string)) {
>>>>>... |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.