method
stringlengths 13
441k
| clean_method
stringlengths 7
313k
| doc
stringlengths 17
17.3k
| comment
stringlengths 3
1.42k
| method_name
stringlengths 1
273
| extra
dict | imports
list | imports_info
stringlengths 19
34.8k
| cluster_imports_info
stringlengths 15
3.66k
| libraries
list | libraries_info
stringlengths 6
661
| id
int64 0
2.92M
|
|---|---|---|---|---|---|---|---|---|---|---|---|
public void print(String string) throws IOException {
if (isBufferStreamBased()) {
String encoding = getContentTypeEncoding();
byte[] bytes;
if (encoding == null) {
bytes = string.getBytes();
} else {
bytes = string.getBytes(encoding);
}
buffer.getOutputStream().write(bytes);
return;
}
// make sure at least writer is initialized
buffer.getWriter().write(string);
}
|
void function(String string) throws IOException { if (isBufferStreamBased()) { String encoding = getContentTypeEncoding(); byte[] bytes; if (encoding == null) { bytes = string.getBytes(); } else { bytes = string.getBytes(encoding); } buffer.getOutputStream().write(bytes); return; } buffer.getWriter().write(string); }
|
/**
* Appends string to the buffer.
*/
|
Appends string to the buffer
|
print
|
{
"repo_name": "007slm/jodd",
"path": "jodd-servlet/src/main/java/jodd/servlet/wrapper/BufferResponseWrapper.java",
"license": "bsd-3-clause",
"size": 12471
}
|
[
"java.io.IOException"
] |
import java.io.IOException;
|
import java.io.*;
|
[
"java.io"
] |
java.io;
| 2,288,381
|
@Deprecated
public void setEndpoint(String endpoint) throws JMSException {
try {
amazonSQSClient.setEndpoint(endpoint);
} catch (IllegalArgumentException ase) {
JMSException jmsException = new JMSException(ase.getMessage());
throw (JMSException) jmsException.initCause(ase);
}
}
|
void function(String endpoint) throws JMSException { try { amazonSQSClient.setEndpoint(endpoint); } catch (IllegalArgumentException ase) { JMSException jmsException = new JMSException(ase.getMessage()); throw (JMSException) jmsException.initCause(ase); } }
|
/**
* Sets SQS endpoint and wraps IllegalArgumentException.
* Deprecated. Instead of manipulating settings of existing AmazonSQS client, provide correct configuration when creating it through SQSConnectionFactory constructors.
*
* @param endpoint
* The endpoint (ex: "sqs.us-east-1.amazonaws.com") of the region
* specific AWS endpoint this client will communicate with.
* @throws JMSException
*/
|
Sets SQS endpoint and wraps IllegalArgumentException. Deprecated. Instead of manipulating settings of existing AmazonSQS client, provide correct configuration when creating it through SQSConnectionFactory constructors
|
setEndpoint
|
{
"repo_name": "awslabs/amazon-sqs-java-messaging-lib",
"path": "src/main/java/com/amazon/sqs/javamessaging/AmazonSQSMessagingClientWrapper.java",
"license": "apache-2.0",
"size": 19435
}
|
[
"javax.jms.JMSException"
] |
import javax.jms.JMSException;
|
import javax.jms.*;
|
[
"javax.jms"
] |
javax.jms;
| 713,632
|
public com.mozu.api.contracts.core.User getUser(String userId, String responseFields) throws Exception
{
MozuClient<com.mozu.api.contracts.core.User> client = com.mozu.api.clients.platform.adminuser.AdminUserClient.getUserClient( userId, responseFields);
client.setContext(_apiContext);
client.executeRequest();
return client.getResult();
}
|
com.mozu.api.contracts.core.User function(String userId, String responseFields) throws Exception { MozuClient<com.mozu.api.contracts.core.User> client = com.mozu.api.clients.platform.adminuser.AdminUserClient.getUserClient( userId, responseFields); client.setContext(_apiContext); client.executeRequest(); return client.getResult(); }
|
/**
* Retrieves the details of the specified administrator user account.
* <p><pre><code>
* AdminUser adminuser = new AdminUser();
* User user = adminuser.getUser( userId, responseFields);
* </code></pre></p>
* @param responseFields Use this field to include those fields which are not included by default.
* @param userId Unique identifier of the user whose tenant scopes you want to retrieve.
* @return com.mozu.api.contracts.core.User
* @see com.mozu.api.contracts.core.User
*/
|
Retrieves the details of the specified administrator user account. <code><code> AdminUser adminuser = new AdminUser(); User user = adminuser.getUser( userId, responseFields); </code></code>
|
getUser
|
{
"repo_name": "bhewett/mozu-java",
"path": "mozu-javaasync-core/src/main/java/com/mozu/api/resources/platform/adminuser/AdminUserResource.java",
"license": "mit",
"size": 7781
}
|
[
"com.mozu.api.MozuClient"
] |
import com.mozu.api.MozuClient;
|
import com.mozu.api.*;
|
[
"com.mozu.api"
] |
com.mozu.api;
| 1,870,959
|
@SuppressWarnings("unchecked")
public BinaryVector copy() {
BinaryVector copy = new BinaryVector(dimension);
copy.bitSet = (FixedBitSet) bitSet.clone();
if (!isSparse)
copy.votingRecord = (ArrayList<FixedBitSet>) votingRecord.clone();
return copy;
}
|
@SuppressWarnings(STR) BinaryVector function() { BinaryVector copy = new BinaryVector(dimension); copy.bitSet = (FixedBitSet) bitSet.clone(); if (!isSparse) copy.votingRecord = (ArrayList<FixedBitSet>) votingRecord.clone(); return copy; }
|
/**
* Returns a new copy of this vector, in dense format.
*/
|
Returns a new copy of this vector, in dense format
|
copy
|
{
"repo_name": "Lucky-Dhakad/semanticvectors",
"path": "src/main/java/pitt/search/semanticvectors/vectors/BinaryVector.java",
"license": "bsd-3-clause",
"size": 31043
}
|
[
"java.util.ArrayList",
"org.apache.lucene.util.FixedBitSet"
] |
import java.util.ArrayList; import org.apache.lucene.util.FixedBitSet;
|
import java.util.*; import org.apache.lucene.util.*;
|
[
"java.util",
"org.apache.lucene"
] |
java.util; org.apache.lucene;
| 2,835,613
|
@Override()
public java.lang.Class getJavaClass(
) {
return org.chocolate_milk.model.FixedAssetSalesInfo.class;
}
|
@Override() java.lang.Class function( ) { return org.chocolate_milk.model.FixedAssetSalesInfo.class; }
|
/**
* Method getJavaClass.
*
* @return the Java class represented by this descriptor.
*/
|
Method getJavaClass
|
getJavaClass
|
{
"repo_name": "galleon1/chocolate-milk",
"path": "src/org/chocolate_milk/model/descriptors/FixedAssetSalesInfoDescriptor.java",
"license": "lgpl-3.0",
"size": 11863
}
|
[
"org.chocolate_milk.model.FixedAssetSalesInfo"
] |
import org.chocolate_milk.model.FixedAssetSalesInfo;
|
import org.chocolate_milk.model.*;
|
[
"org.chocolate_milk.model"
] |
org.chocolate_milk.model;
| 2,748,229
|
public Observable<ServiceResponse<TroubleshootingResultInner>> beginGetTroubleshootingWithServiceResponseAsync(String resourceGroupName, String networkWatcherName, TroubleshootingParameters parameters) {
if (resourceGroupName == null) {
throw new IllegalArgumentException("Parameter resourceGroupName is required and cannot be null.");
}
if (networkWatcherName == null) {
throw new IllegalArgumentException("Parameter networkWatcherName is required and cannot be null.");
}
if (this.client.subscriptionId() == null) {
throw new IllegalArgumentException("Parameter this.client.subscriptionId() is required and cannot be null.");
}
if (parameters == null) {
throw new IllegalArgumentException("Parameter parameters is required and cannot be null.");
}
|
Observable<ServiceResponse<TroubleshootingResultInner>> function(String resourceGroupName, String networkWatcherName, TroubleshootingParameters parameters) { if (resourceGroupName == null) { throw new IllegalArgumentException(STR); } if (networkWatcherName == null) { throw new IllegalArgumentException(STR); } if (this.client.subscriptionId() == null) { throw new IllegalArgumentException(STR); } if (parameters == null) { throw new IllegalArgumentException(STR); }
|
/**
* Initiate troubleshooting on a specified resource.
*
* @param resourceGroupName The name of the resource group.
* @param networkWatcherName The name of the network watcher resource.
* @param parameters Parameters that define the resource to troubleshoot.
* @throws IllegalArgumentException thrown if parameters fail the validation
* @return the observable to the TroubleshootingResultInner object
*/
|
Initiate troubleshooting on a specified resource
|
beginGetTroubleshootingWithServiceResponseAsync
|
{
"repo_name": "selvasingh/azure-sdk-for-java",
"path": "sdk/network/mgmt-v2020_03_01/src/main/java/com/microsoft/azure/management/network/v2020_03_01/implementation/NetworkWatchersInner.java",
"license": "mit",
"size": 190989
}
|
[
"com.microsoft.azure.management.network.v2020_03_01.TroubleshootingParameters",
"com.microsoft.rest.ServiceResponse"
] |
import com.microsoft.azure.management.network.v2020_03_01.TroubleshootingParameters; import com.microsoft.rest.ServiceResponse;
|
import com.microsoft.azure.management.network.v2020_03_01.*; import com.microsoft.rest.*;
|
[
"com.microsoft.azure",
"com.microsoft.rest"
] |
com.microsoft.azure; com.microsoft.rest;
| 2,227,879
|
public void incrementCount(List<K> l) {
incrementCount(l, 1.0);
}
|
void function(List<K> l) { incrementCount(l, 1.0); }
|
/**
* Equivalent to incrementCount(l, 1.0).
*/
|
Equivalent to incrementCount(l, 1.0)
|
incrementCount
|
{
"repo_name": "heeyounglee/hcoref",
"path": "src/edu/stanford/nlp/stats/GeneralizedCounter.java",
"license": "gpl-2.0",
"size": 29230
}
|
[
"java.util.List"
] |
import java.util.List;
|
import java.util.*;
|
[
"java.util"
] |
java.util;
| 1,145,269
|
public double computeEffectiveParticles(
DataDistribution<ParameterType> particles );
public static interface Updater<ObservationType,ParameterType>
extends CloneableSerializable
{
|
double function( DataDistribution<ParameterType> particles ); public static interface Updater<ObservationType,ParameterType> extends CloneableSerializable {
|
/**
* Computes the effective number of particles.
* @param particles
* Current state of the Particle filter.
* @return
* Effective number of particles.
*/
|
Computes the effective number of particles
|
computeEffectiveParticles
|
{
"repo_name": "codeaudit/Foundry",
"path": "Components/LearningCore/Source/gov/sandia/cognition/statistics/bayesian/ParticleFilter.java",
"license": "bsd-3-clause",
"size": 4514
}
|
[
"gov.sandia.cognition.statistics.DataDistribution",
"gov.sandia.cognition.util.CloneableSerializable"
] |
import gov.sandia.cognition.statistics.DataDistribution; import gov.sandia.cognition.util.CloneableSerializable;
|
import gov.sandia.cognition.statistics.*; import gov.sandia.cognition.util.*;
|
[
"gov.sandia.cognition"
] |
gov.sandia.cognition;
| 1,418,762
|
if (t.isClass()) {
return JInterfaceType.createUnresolvableType();
}
Preconditions.checkArgument(t.isInterface());
String typeName = NameConverter.convertClassOrInterfaceToFullName(t);
if (scope.containsInterfaceType(typeName)) {
return scope.getInterfaceType(typeName);
} else {
return scope.createNewInterfaceType(t);
}
}
|
if (t.isClass()) { return JInterfaceType.createUnresolvableType(); } Preconditions.checkArgument(t.isInterface()); String typeName = NameConverter.convertClassOrInterfaceToFullName(t); if (scope.containsInterfaceType(typeName)) { return scope.getInterfaceType(typeName); } else { return scope.createNewInterfaceType(t); } }
|
/**
* Searches for a type within the Type Hierarchy.
* If found, returns it.
*
* @param t binding representing the sought after type.
* @return Returns a type within the TypeHierachie or a Unspecified Type.
*/
|
Searches for a type within the Type Hierarchy. If found, returns it
|
convertInterfaceType
|
{
"repo_name": "nishanttotla/predator",
"path": "cpachecker/src/org/sosy_lab/cpachecker/cfa/parser/eclipse/java/ASTTypeConverter.java",
"license": "gpl-3.0",
"size": 2647
}
|
[
"com.google.common.base.Preconditions",
"org.sosy_lab.cpachecker.cfa.types.java.JInterfaceType"
] |
import com.google.common.base.Preconditions; import org.sosy_lab.cpachecker.cfa.types.java.JInterfaceType;
|
import com.google.common.base.*; import org.sosy_lab.cpachecker.cfa.types.java.*;
|
[
"com.google.common",
"org.sosy_lab.cpachecker"
] |
com.google.common; org.sosy_lab.cpachecker;
| 2,636,734
|
public static <T> Iterator<T> iterator( Enumeration<T> e )
{
return new Iterator<T>()
{
|
static <T> Iterator<T> function( Enumeration<T> e ) { return new Iterator<T>() {
|
/**
* Wrap a legacy {@link Enumeration} into an Iterator
* <p>
* @param <T>
* @param e
* <p>
* @return
*/
|
Wrap a legacy <code>Enumeration</code> into an Iterator
|
iterator
|
{
"repo_name": "peter-mount/opendata-common",
"path": "core/src/main/java/uk/trainwatch/util/CollectionUtils.java",
"license": "apache-2.0",
"size": 15855
}
|
[
"java.util.Enumeration",
"java.util.Iterator"
] |
import java.util.Enumeration; import java.util.Iterator;
|
import java.util.*;
|
[
"java.util"
] |
java.util;
| 2,658,354
|
@SuppressWarnings("unchecked")
public List<String> getObservationIdentifier(Session session) {
Criteria criteria =
session.createCriteria(getObservationFactory().contextualReferencedClass())
.add(Restrictions.eq(ContextualReferencedObservation.DELETED, false))
.add(Restrictions.isNotNull(ContextualReferencedObservation.IDENTIFIER))
.setProjection(Projections.distinct(Projections.property(ContextualReferencedObservation.IDENTIFIER)));
LOGGER.debug("QUERY getObservationIdentifiers(): {}", HibernateHelper.getSqlString(criteria));
return criteria.list();
}
|
@SuppressWarnings(STR) List<String> function(Session session) { Criteria criteria = session.createCriteria(getObservationFactory().contextualReferencedClass()) .add(Restrictions.eq(ContextualReferencedObservation.DELETED, false)) .add(Restrictions.isNotNull(ContextualReferencedObservation.IDENTIFIER)) .setProjection(Projections.distinct(Projections.property(ContextualReferencedObservation.IDENTIFIER))); LOGGER.debug(STR, HibernateHelper.getSqlString(criteria)); return criteria.list(); }
|
/**
* Get all observation identifiers
*
* @param session
* Hibernate session
* @return Observation identifiers
*/
|
Get all observation identifiers
|
getObservationIdentifier
|
{
"repo_name": "ahuarte47/SOS",
"path": "hibernate/common/src/main/java/org/n52/sos/ds/hibernate/dao/observation/AbstractObservationDAO.java",
"license": "gpl-2.0",
"size": 82385
}
|
[
"java.util.List",
"org.hibernate.Criteria",
"org.hibernate.Session",
"org.hibernate.criterion.Projections",
"org.hibernate.criterion.Restrictions",
"org.n52.sos.ds.hibernate.entities.observation.ContextualReferencedObservation",
"org.n52.sos.ds.hibernate.util.HibernateHelper"
] |
import java.util.List; import org.hibernate.Criteria; import org.hibernate.Session; import org.hibernate.criterion.Projections; import org.hibernate.criterion.Restrictions; import org.n52.sos.ds.hibernate.entities.observation.ContextualReferencedObservation; import org.n52.sos.ds.hibernate.util.HibernateHelper;
|
import java.util.*; import org.hibernate.*; import org.hibernate.criterion.*; import org.n52.sos.ds.hibernate.entities.observation.*; import org.n52.sos.ds.hibernate.util.*;
|
[
"java.util",
"org.hibernate",
"org.hibernate.criterion",
"org.n52.sos"
] |
java.util; org.hibernate; org.hibernate.criterion; org.n52.sos;
| 1,773,914
|
public byte[] getBytesFromTextChild() throws XMLSecurityException {
return Base64.decode(XMLUtils.getFullTextChildrenFromElement(this.constructionElement));
}
|
byte[] function() throws XMLSecurityException { return Base64.decode(XMLUtils.getFullTextChildrenFromElement(this.constructionElement)); }
|
/**
* Method getBytesFromTextChild
*
* @return The base64 bytes from the text children of this element
* @throws XMLSecurityException
*/
|
Method getBytesFromTextChild
|
getBytesFromTextChild
|
{
"repo_name": "rokn/Count_Words_2015",
"path": "testing/openjdk2/jdk/src/share/classes/com/sun/org/apache/xml/internal/security/utils/ElementProxy.java",
"license": "mit",
"size": 16335
}
|
[
"com.sun.org.apache.xml.internal.security.exceptions.XMLSecurityException"
] |
import com.sun.org.apache.xml.internal.security.exceptions.XMLSecurityException;
|
import com.sun.org.apache.xml.internal.security.exceptions.*;
|
[
"com.sun.org"
] |
com.sun.org;
| 130,348
|
@Override
protected Instances processDatasetWithClusterer(Instances data, Clusterer clusterer) {
Instances result;
Map<Integer,Integer> counts;
int i;
int cluster;
List<Integer> clusters;
DenseInstance inst;
result = createOutputFormat(data);
counts = new HashMap<>();
// cluster instances
for (i = 0; i < data.numInstances(); i++) {
try {
cluster = clusterer.clusterInstance(data.instance(i));
if (!counts.containsKey(cluster))
counts.put(cluster, 0);
counts.put(cluster, counts.get(cluster) + 1);
}
catch (Exception e) {
getLogger().log(Level.SEVERE, "Failed to cluster instance #" + (i+1) + "!", e);
}
}
// generate overview
clusters = new ArrayList<>(counts.keySet());
Collections.sort(clusters);
for (i = 0; i < clusters.size(); i++) {
cluster = clusters.get(i);
inst = new DenseInstance(1.0, new double[]{cluster, counts.get(cluster)});
result.add(inst);
}
return result;
}
|
Instances function(Instances data, Clusterer clusterer) { Instances result; Map<Integer,Integer> counts; int i; int cluster; List<Integer> clusters; DenseInstance inst; result = createOutputFormat(data); counts = new HashMap<>(); for (i = 0; i < data.numInstances(); i++) { try { cluster = clusterer.clusterInstance(data.instance(i)); if (!counts.containsKey(cluster)) counts.put(cluster, 0); counts.put(cluster, counts.get(cluster) + 1); } catch (Exception e) { getLogger().log(Level.SEVERE, STR + (i+1) + "!", e); } } clusters = new ArrayList<>(counts.keySet()); Collections.sort(clusters); for (i = 0; i < clusters.size(); i++) { cluster = clusters.get(i); inst = new DenseInstance(1.0, new double[]{cluster, counts.get(cluster)}); result.add(inst); } return result; }
|
/**
* Performs some form of processing on the full dataset.
*/
|
Performs some form of processing on the full dataset
|
processDatasetWithClusterer
|
{
"repo_name": "waikato-datamining/adams-base",
"path": "adams-weka/src/main/java/adams/flow/transformer/wekaclusterer/ClusterCounts.java",
"license": "gpl-3.0",
"size": 3770
}
|
[
"java.util.ArrayList",
"java.util.Collections",
"java.util.HashMap",
"java.util.List",
"java.util.Map",
"java.util.logging.Level"
] |
import java.util.ArrayList; import java.util.Collections; import java.util.HashMap; import java.util.List; import java.util.Map; import java.util.logging.Level;
|
import java.util.*; import java.util.logging.*;
|
[
"java.util"
] |
java.util;
| 2,366,466
|
public CmsSimplePropertyEditor getPropertyEditor() {
return m_propertyEditor;
}
|
CmsSimplePropertyEditor function() { return m_propertyEditor; }
|
/**
* Gets the property editor instance.<p>
*
* @return the property editor instance
*/
|
Gets the property editor instance
|
getPropertyEditor
|
{
"repo_name": "it-tavis/opencms-core",
"path": "src-gwt/org/opencms/ade/postupload/client/ui/CmsUploadPropertyPanel.java",
"license": "lgpl-2.1",
"size": 6498
}
|
[
"org.opencms.gwt.client.property.CmsSimplePropertyEditor"
] |
import org.opencms.gwt.client.property.CmsSimplePropertyEditor;
|
import org.opencms.gwt.client.property.*;
|
[
"org.opencms.gwt"
] |
org.opencms.gwt;
| 1,324,365
|
@Test
public void testServerRestartWithNewTypes() throws Exception {
IgniteEx node1 = start(1, KeyClass.class, ValueClass.class);
assertTypes(node1, ValueClass.class);
IgniteEx node2 = startClientNoCache(2);
GridCacheContext<Object, Object> context0 = node2.context().cache().context().cacheContext(CU.cacheId(CACHE_NAME));
node2.cache(CACHE_NAME);
GridCacheContext<Object, Object> context = node2.context().cache().context().cacheContext(CU.cacheId(CACHE_NAME));
GridCacheAdapter<Object, Object> entries = node2.context().cache().internalCache(CACHE_NAME);
assertTrue(entries.active());
node2.cache(CACHE_NAME);
assertTypes(node2, ValueClass.class);
stopGrid(1);
|
void function() throws Exception { IgniteEx node1 = start(1, KeyClass.class, ValueClass.class); assertTypes(node1, ValueClass.class); IgniteEx node2 = startClientNoCache(2); GridCacheContext<Object, Object> context0 = node2.context().cache().context().cacheContext(CU.cacheId(CACHE_NAME)); node2.cache(CACHE_NAME); GridCacheContext<Object, Object> context = node2.context().cache().context().cacheContext(CU.cacheId(CACHE_NAME)); GridCacheAdapter<Object, Object> entries = node2.context().cache().internalCache(CACHE_NAME); assertTrue(entries.active()); node2.cache(CACHE_NAME); assertTypes(node2, ValueClass.class); stopGrid(1);
|
/**
* Test client reconnect after server restart accompanied by schema change.
*
* @throws Exception If failed.
*/
|
Test client reconnect after server restart accompanied by schema change
|
testServerRestartWithNewTypes
|
{
"repo_name": "NSAmelchev/ignite",
"path": "modules/indexing/src/test/java/org/apache/ignite/internal/processors/cache/index/SchemaExchangeSelfTest.java",
"license": "apache-2.0",
"size": 21252
}
|
[
"org.apache.ignite.internal.IgniteEx",
"org.apache.ignite.internal.processors.cache.GridCacheAdapter",
"org.apache.ignite.internal.processors.cache.GridCacheContext",
"org.apache.ignite.internal.util.typedef.internal.CU"
] |
import org.apache.ignite.internal.IgniteEx; import org.apache.ignite.internal.processors.cache.GridCacheAdapter; import org.apache.ignite.internal.processors.cache.GridCacheContext; import org.apache.ignite.internal.util.typedef.internal.CU;
|
import org.apache.ignite.internal.*; import org.apache.ignite.internal.processors.cache.*; import org.apache.ignite.internal.util.typedef.internal.*;
|
[
"org.apache.ignite"
] |
org.apache.ignite;
| 1,235,059
|
public void setValueAnimated(float _valueTo, long _animationDuration) {
mAnimationDuration = _animationDuration;
Message msg = new Message();
msg.what = AnimationMsg.SET_VALUE_ANIMATED.ordinal();
msg.obj = new float[]{mCurrentValue, _valueTo};
mAnimationHandler.sendMessage(msg);
}
|
void function(float _valueTo, long _animationDuration) { mAnimationDuration = _animationDuration; Message msg = new Message(); msg.what = AnimationMsg.SET_VALUE_ANIMATED.ordinal(); msg.obj = new float[]{mCurrentValue, _valueTo}; mAnimationHandler.sendMessage(msg); }
|
/**
* Sets the value of the circle view with an animation.
* The current value is used as the start value of the animation
*
* @param _valueTo value after animation
* @param _animationDuration the duration of the animation in milliseconds.
*/
|
Sets the value of the circle view with an animation. The current value is used as the start value of the animation
|
setValueAnimated
|
{
"repo_name": "captainbupt/Circle-Progress-View",
"path": "CircleProgressView/src/main/java/at/grabner/circleprogress/CircleProgressView.java",
"license": "mit",
"size": 52578
}
|
[
"android.os.Message"
] |
import android.os.Message;
|
import android.os.*;
|
[
"android.os"
] |
android.os;
| 1,427,121
|
@Test
public void testT1RV6D5_T1LV2D3() {
test_id = getTestId("T1RV6D5", "T1LV2D3", "225");
String src = selectTRVD("T1RV6D5");
String dest = selectTLVD("T1LV2D3");
String result = ".";
try {
result = TRVD_TLVD_Action(src, dest);
} catch (RecognitionException e) {
e.printStackTrace();
} catch (TokenStreamException e) {
e.printStackTrace();
}
assertTrue(Failure2, checkResult_Failure2(src, dest, result));
GraphicalEditor editor = getActiveEditor();
if (editor != null) {
validateOrGenerateResults(editor, generateResults);
}
}
|
void function() { test_id = getTestId(STR, STR, "225"); String src = selectTRVD(STR); String dest = selectTLVD(STR); String result = "."; try { result = TRVD_TLVD_Action(src, dest); } catch (RecognitionException e) { e.printStackTrace(); } catch (TokenStreamException e) { e.printStackTrace(); } assertTrue(Failure2, checkResult_Failure2(src, dest, result)); GraphicalEditor editor = getActiveEditor(); if (editor != null) { validateOrGenerateResults(editor, generateResults); } }
|
/**
* Perform the test for the given matrix column (T1RV6D5) and row (T1LV2D3).
*
*/
|
Perform the test for the given matrix column (T1RV6D5) and row (T1LV2D3)
|
testT1RV6D5_T1LV2D3
|
{
"repo_name": "jason-rhodes/bridgepoint",
"path": "src/org.xtuml.bp.als.oal.test/src/org/xtuml/bp/als/oal/test/SingleDimensionFixedArrayAssigmentTest_12_Generics.java",
"license": "apache-2.0",
"size": 155634
}
|
[
"org.xtuml.bp.ui.graphics.editor.GraphicalEditor"
] |
import org.xtuml.bp.ui.graphics.editor.GraphicalEditor;
|
import org.xtuml.bp.ui.graphics.editor.*;
|
[
"org.xtuml.bp"
] |
org.xtuml.bp;
| 1,489,505
|
public ActivityPolicy withAdditionalProperties(Map<String, Object> additionalProperties) {
this.additionalProperties = additionalProperties;
return this;
}
|
ActivityPolicy function(Map<String, Object> additionalProperties) { this.additionalProperties = additionalProperties; return this; }
|
/**
* Set the additionalProperties property: Execution policy for an activity.
*
* @param additionalProperties the additionalProperties value to set.
* @return the ActivityPolicy object itself.
*/
|
Set the additionalProperties property: Execution policy for an activity
|
withAdditionalProperties
|
{
"repo_name": "Azure/azure-sdk-for-java",
"path": "sdk/datafactory/azure-resourcemanager-datafactory/src/main/java/com/azure/resourcemanager/datafactory/models/ActivityPolicy.java",
"license": "mit",
"size": 6654
}
|
[
"java.util.Map"
] |
import java.util.Map;
|
import java.util.*;
|
[
"java.util"
] |
java.util;
| 1,125,768
|
public Pair<Double, INDArray> nn(INDArray point) {
return nn(root, point, rect, Double.POSITIVE_INFINITY, null, 0);
}
|
Pair<Double, INDArray> function(INDArray point) { return nn(root, point, rect, Double.POSITIVE_INFINITY, null, 0); }
|
/**
* Query for nearest neighbor. Returns the distance and point
* @param point the point to query for
* @return
*/
|
Query for nearest neighbor. Returns the distance and point
|
nn
|
{
"repo_name": "RobAltena/deeplearning4j",
"path": "deeplearning4j/deeplearning4j-nearestneighbors-parent/nearestneighbor-core/src/main/java/org/deeplearning4j/clustering/kdtree/KDTree.java",
"license": "apache-2.0",
"size": 12115
}
|
[
"org.nd4j.linalg.api.ndarray.INDArray",
"org.nd4j.linalg.primitives.Pair"
] |
import org.nd4j.linalg.api.ndarray.INDArray; import org.nd4j.linalg.primitives.Pair;
|
import org.nd4j.linalg.api.ndarray.*; import org.nd4j.linalg.primitives.*;
|
[
"org.nd4j.linalg"
] |
org.nd4j.linalg;
| 2,492,444
|
@Test
public void getFixListDefault() throws P4JavaException {
when(server.execMapCmdList(eq(FIXES.toString()),
argThat(FIX_FS_PARAMS_UNKNOWN_MATCHER), eq(null)))
.thenReturn(buildValidResultMap());
List<IFileSpec> specs = FileSpecBuilder.makeFileSpecList(DEPOT_DEV_PATH);
List<IFix> fixes = fixesDelegator.getFixList(specs, Changelist.DEFAULT,
TEST_JOB_123, true, 0);
verify(server).execMapCmdList(eq(FIXES.toString()), argThat(FIX_FS_PARAMS_UNKNOWN_MATCHER),
eq(null));
assertFixes(fixes);
}
|
void function() throws P4JavaException { when(server.execMapCmdList(eq(FIXES.toString()), argThat(FIX_FS_PARAMS_UNKNOWN_MATCHER), eq(null))) .thenReturn(buildValidResultMap()); List<IFileSpec> specs = FileSpecBuilder.makeFileSpecList(DEPOT_DEV_PATH); List<IFix> fixes = fixesDelegator.getFixList(specs, Changelist.DEFAULT, TEST_JOB_123, true, 0); verify(server).execMapCmdList(eq(FIXES.toString()), argThat(FIX_FS_PARAMS_UNKNOWN_MATCHER), eq(null)); assertFixes(fixes); }
|
/**
* Gets the fix list for a default.
*
* @throws P4JavaException
* the p4 java exception
*/
|
Gets the fix list for a default
|
getFixListDefault
|
{
"repo_name": "groboclown/p4ic4idea",
"path": "p4java/r18-1/src/test/java/com/perforce/p4java/impl/mapbased/server/cmd/FixesDelegatorTest.java",
"license": "apache-2.0",
"size": 14139
}
|
[
"com.perforce.p4java.core.IFix",
"com.perforce.p4java.core.file.FileSpecBuilder",
"com.perforce.p4java.core.file.IFileSpec",
"com.perforce.p4java.exception.P4JavaException",
"com.perforce.p4java.impl.generic.core.Changelist",
"com.perforce.p4java.server.CmdSpec",
"java.util.List",
"org.mockito.ArgumentMatchers",
"org.mockito.Mockito"
] |
import com.perforce.p4java.core.IFix; import com.perforce.p4java.core.file.FileSpecBuilder; import com.perforce.p4java.core.file.IFileSpec; import com.perforce.p4java.exception.P4JavaException; import com.perforce.p4java.impl.generic.core.Changelist; import com.perforce.p4java.server.CmdSpec; import java.util.List; import org.mockito.ArgumentMatchers; import org.mockito.Mockito;
|
import com.perforce.p4java.core.*; import com.perforce.p4java.core.file.*; import com.perforce.p4java.exception.*; import com.perforce.p4java.impl.generic.core.*; import com.perforce.p4java.server.*; import java.util.*; import org.mockito.*;
|
[
"com.perforce.p4java",
"java.util",
"org.mockito"
] |
com.perforce.p4java; java.util; org.mockito;
| 320,981
|
private void logSlowRequests(long requestLifetime, Request<?> request,
byte[] responseContents, StatusLine statusLine) {
if (DEBUG || requestLifetime > SLOW_REQUEST_THRESHOLD_MS) {
VolleyLog.d("HTTP response for request=<%s> [lifetime=%d], [size=%s], " +
"[rc=%d], [retryCount=%s]", request, requestLifetime,
responseContents != null ? responseContents.length : "null",
statusLine.getStatusCode(), request.getRetryPolicy().getCurrentRetryCount());
}
}
|
void function(long requestLifetime, Request<?> request, byte[] responseContents, StatusLine statusLine) { if (DEBUG requestLifetime > SLOW_REQUEST_THRESHOLD_MS) { VolleyLog.d(STR + STR, request, requestLifetime, responseContents != null ? responseContents.length : "null", statusLine.getStatusCode(), request.getRetryPolicy().getCurrentRetryCount()); } }
|
/**
* Logs requests that took over SLOW_REQUEST_THRESHOLD_MS to complete.
*/
|
Logs requests that took over SLOW_REQUEST_THRESHOLD_MS to complete
|
logSlowRequests
|
{
"repo_name": "shazangroup/Mobograph",
"path": "TMessagesProj/src/main/java/org/telegram/messenger/volley/toolbox/BasicNetwork.java",
"license": "gpl-2.0",
"size": 11099
}
|
[
"org.apache.http.StatusLine",
"org.telegram.messenger.volley.Request",
"org.telegram.messenger.volley.VolleyLog"
] |
import org.apache.http.StatusLine; import org.telegram.messenger.volley.Request; import org.telegram.messenger.volley.VolleyLog;
|
import org.apache.http.*; import org.telegram.messenger.volley.*;
|
[
"org.apache.http",
"org.telegram.messenger"
] |
org.apache.http; org.telegram.messenger;
| 909,494
|
@Override
public void getFeature(Point request, StreamObserver<Feature> responseObserver) {
responseObserver.onNext(checkFeature(request));
responseObserver.onCompleted();
}
|
void function(Point request, StreamObserver<Feature> responseObserver) { responseObserver.onNext(checkFeature(request)); responseObserver.onCompleted(); }
|
/**
* Gets the {@link Feature} at the requested {@link Point}. If no feature at that location
* exists, an unnamed feature is returned at the provided location.
*
* @param request the requested location for the feature.
* @param responseObserver the observer that will receive the feature at the requested point.
*/
|
Gets the <code>Feature</code> at the requested <code>Point</code>. If no feature at that location exists, an unnamed feature is returned at the provided location
|
getFeature
|
{
"repo_name": "robinbakkerus/workshop-grpc",
"path": "src/main/java/io/grpc/examples/routeguide/RouteGuideServer.java",
"license": "apache-2.0",
"size": 11416
}
|
[
"io.grpc.stub.StreamObserver"
] |
import io.grpc.stub.StreamObserver;
|
import io.grpc.stub.*;
|
[
"io.grpc.stub"
] |
io.grpc.stub;
| 1,484,894
|
@Override
public Collection<? extends EStructuralFeature> getChildrenFeatures(Object object) {
if (childrenFeatures == null) {
super.getChildrenFeatures(object);
childrenFeatures.add(TurtlebotmissionPackage.Literals.MISSION__TASK);
}
return childrenFeatures;
}
|
Collection<? extends EStructuralFeature> function(Object object) { if (childrenFeatures == null) { super.getChildrenFeatures(object); childrenFeatures.add(TurtlebotmissionPackage.Literals.MISSION__TASK); } return childrenFeatures; }
|
/**
* This specifies how to implement {@link #getChildren} and is used to deduce an appropriate feature for an
* {@link org.eclipse.emf.edit.command.AddCommand}, {@link org.eclipse.emf.edit.command.RemoveCommand} or
* {@link org.eclipse.emf.edit.command.MoveCommand} in {@link #createCommand}.
* <!-- begin-user-doc -->
* <!-- end-user-doc -->
* @generated
*/
|
This specifies how to implement <code>#getChildren</code> and is used to deduce an appropriate feature for an <code>org.eclipse.emf.edit.command.AddCommand</code>, <code>org.eclipse.emf.edit.command.RemoveCommand</code> or <code>org.eclipse.emf.edit.command.MoveCommand</code> in <code>#createCommand</code>.
|
getChildrenFeatures
|
{
"repo_name": "kribe48/wasp-mbse",
"path": "WASP-turtlebot-DSL/CodeFrame/se.chalmers.turtlebotmission.edit/src/turtlebotmission/provider/MissionItemProvider.java",
"license": "mit",
"size": 4644
}
|
[
"java.util.Collection",
"org.eclipse.emf.ecore.EStructuralFeature"
] |
import java.util.Collection; import org.eclipse.emf.ecore.EStructuralFeature;
|
import java.util.*; import org.eclipse.emf.ecore.*;
|
[
"java.util",
"org.eclipse.emf"
] |
java.util; org.eclipse.emf;
| 2,770,749
|
public static MozuClient<com.mozu.api.contracts.commerceruntime.payments.PaymentCollection> getPaymentsClient(String orderId, String responseFields) throws Exception
{
MozuUrl url = com.mozu.api.urls.commerce.orders.PaymentUrl.getPaymentsUrl(orderId, responseFields);
String verb = "GET";
Class<?> clz = com.mozu.api.contracts.commerceruntime.payments.PaymentCollection.class;
MozuClient<com.mozu.api.contracts.commerceruntime.payments.PaymentCollection> mozuClient = (MozuClient<com.mozu.api.contracts.commerceruntime.payments.PaymentCollection>) MozuClientFactory.getInstance(clz);
mozuClient.setVerb(verb);
mozuClient.setResourceUrl(url);
return mozuClient;
}
|
static MozuClient<com.mozu.api.contracts.commerceruntime.payments.PaymentCollection> function(String orderId, String responseFields) throws Exception { MozuUrl url = com.mozu.api.urls.commerce.orders.PaymentUrl.getPaymentsUrl(orderId, responseFields); String verb = "GET"; Class<?> clz = com.mozu.api.contracts.commerceruntime.payments.PaymentCollection.class; MozuClient<com.mozu.api.contracts.commerceruntime.payments.PaymentCollection> mozuClient = (MozuClient<com.mozu.api.contracts.commerceruntime.payments.PaymentCollection>) MozuClientFactory.getInstance(clz); mozuClient.setVerb(verb); mozuClient.setResourceUrl(url); return mozuClient; }
|
/**
* Retrieves information about all payment transactions submitted for the specified order.
* <p><pre><code>
* MozuClient<com.mozu.api.contracts.commerceruntime.payments.PaymentCollection> mozuClient=GetPaymentsClient( orderId, responseFields);
* client.setBaseAddress(url);
* client.executeRequest();
* PaymentCollection paymentCollection = client.Result();
* </code></pre></p>
* @param orderId Unique identifier of the order.
* @param responseFields Use this field to include those fields which are not included by default.
* @return Mozu.Api.MozuClient <com.mozu.api.contracts.commerceruntime.payments.PaymentCollection>
* @see com.mozu.api.contracts.commerceruntime.payments.PaymentCollection
*/
|
Retrieves information about all payment transactions submitted for the specified order. <code><code> MozuClient mozuClient=GetPaymentsClient( orderId, responseFields); client.setBaseAddress(url); client.executeRequest(); PaymentCollection paymentCollection = client.Result(); </code></code>
|
getPaymentsClient
|
{
"repo_name": "johngatti/mozu-java",
"path": "mozu-java-core/src/main/java/com/mozu/api/clients/commerce/orders/PaymentClient.java",
"license": "mit",
"size": 12133
}
|
[
"com.mozu.api.MozuClient",
"com.mozu.api.MozuClientFactory",
"com.mozu.api.MozuUrl"
] |
import com.mozu.api.MozuClient; import com.mozu.api.MozuClientFactory; import com.mozu.api.MozuUrl;
|
import com.mozu.api.*;
|
[
"com.mozu.api"
] |
com.mozu.api;
| 625,904
|
@Test
public void checkFailedPingToEntityError() throws Exception {
ThreadedDummyConnection threadedCon = getAuthentiactedDummyConnection();
//@formatter:off
String reply =
"<iq type='error' id='qrzSp-16' to='test@myserver.com'>" +
"<ping xmlns='urn:xmpp:ping'/>" +
"<error type='cancel'>" +
"<service-unavailable xmlns='urn:ietf:params:xml:ns:xmpp-stanzas'/>" +
"</error>" +
"</iq>";
//@formatter:on
IQ serviceUnavailable = (IQ) PacketParserUtils.parseStanza(reply);
threadedCon.addIQReply(serviceUnavailable);
PingManager pinger = PingManager.getInstanceFor(threadedCon);
boolean pingSuccess = pinger.ping(DUMMY_AT_EXAMPLE_ORG);
assertFalse(pingSuccess);
}
|
void function() throws Exception { ThreadedDummyConnection threadedCon = getAuthentiactedDummyConnection(); String reply = STR + STR + STR + STR + STR + "</iq>"; IQ serviceUnavailable = (IQ) PacketParserUtils.parseStanza(reply); threadedCon.addIQReply(serviceUnavailable); PingManager pinger = PingManager.getInstanceFor(threadedCon); boolean pingSuccess = pinger.ping(DUMMY_AT_EXAMPLE_ORG); assertFalse(pingSuccess); }
|
/**
* Server returns an exception for entity.
* @throws Exception
*/
|
Server returns an exception for entity
|
checkFailedPingToEntityError
|
{
"repo_name": "esl/Smack",
"path": "smack-extensions/src/test/java/org/jivesoftware/smackx/ping/PingTest.java",
"license": "apache-2.0",
"size": 9386
}
|
[
"org.jivesoftware.smack.ThreadedDummyConnection",
"org.jivesoftware.smack.util.PacketParserUtils",
"org.junit.Assert"
] |
import org.jivesoftware.smack.ThreadedDummyConnection; import org.jivesoftware.smack.util.PacketParserUtils; import org.junit.Assert;
|
import org.jivesoftware.smack.*; import org.jivesoftware.smack.util.*; import org.junit.*;
|
[
"org.jivesoftware.smack",
"org.junit"
] |
org.jivesoftware.smack; org.junit;
| 1,881,640
|
public void setColor(Color color);
|
void function(Color color);
|
/**
* Postavlja boju valnog oblika
*
* @param color zeljena boja
*/
|
Postavlja boju valnog oblika
|
setColor
|
{
"repo_name": "mbezjak/vhdllab",
"path": "vhdllab-client/src/main/java/hr/fer/zemris/vhdllab/applets/simulations/Shape.java",
"license": "apache-2.0",
"size": 1978
}
|
[
"java.awt.Color"
] |
import java.awt.Color;
|
import java.awt.*;
|
[
"java.awt"
] |
java.awt;
| 2,078,795
|
@Bean
AutoPilot autoPilot() {
return new AutoPilot(timeService(),
imuSensor(),
positionSensor(),
new DefaultPidControl(2,
0.0,
0.0,
PI / 4,
PI / 4),
servoController(),
bankingAngleCalculator());
}
|
AutoPilot autoPilot() { return new AutoPilot(timeService(), imuSensor(), positionSensor(), new DefaultPidControl(2, 0.0, 0.0, PI / 4, PI / 4), servoController(), bankingAngleCalculator()); }
|
/**
* Creates the {@link AutoPilot} bean.
*
* @return The {@link AutoPilot} bean.
*/
|
Creates the <code>AutoPilot</code> bean
|
autoPilot
|
{
"repo_name": "harmenweber/space-project",
"path": "src/main/java/ch/harmen/spaceproject/SpaceProjectApplicationContext.java",
"license": "mit",
"size": 3001
}
|
[
"ch.harmen.iot.DefaultPidControl"
] |
import ch.harmen.iot.DefaultPidControl;
|
import ch.harmen.iot.*;
|
[
"ch.harmen.iot"
] |
ch.harmen.iot;
| 135,104
|
private void initTilesUtil() throws ServletException {
if (TilesUtil.isTilesUtilImplSet()) {
log.debug("Skipping re-init of Tiles Plugin. Values defined in the " +
"first initialized plugin take precedence.");
return;
}
// Check if user has specified a TilesUtil implementation classname or not.
// If no implementation is specified, check if user has specified one
// shared single factory for all module, or one factory for each module.
if (this.getTilesUtilImplClassname() == null) {
if (isModuleAware()) {
TilesUtil.setTilesUtil(new TilesUtilStrutsModulesImpl());
} else {
TilesUtil.setTilesUtil(new TilesUtilStrutsImpl());
}
} else { // A classname is specified for the tilesUtilImp, use it.
try {
TilesUtilStrutsImpl impl =
(TilesUtilStrutsImpl) RequestUtils
.applicationClass(getTilesUtilImplClassname())
.newInstance();
TilesUtil.setTilesUtil(impl);
} catch (ClassCastException ex) {
throw new ServletException(
"Can't set TilesUtil implementation to '"
+ getTilesUtilImplClassname()
+ "'. TilesUtil implementation should be a subclass of '"
+ TilesUtilStrutsImpl.class.getName()
+ "'", ex);
} catch (Exception ex) {
throw new ServletException(
"Can't set TilesUtil implementation.",
ex);
}
}
}
|
void function() throws ServletException { if (TilesUtil.isTilesUtilImplSet()) { log.debug(STR + STR); return; } if (this.getTilesUtilImplClassname() == null) { if (isModuleAware()) { TilesUtil.setTilesUtil(new TilesUtilStrutsModulesImpl()); } else { TilesUtil.setTilesUtil(new TilesUtilStrutsImpl()); } } else { try { TilesUtilStrutsImpl impl = (TilesUtilStrutsImpl) RequestUtils .applicationClass(getTilesUtilImplClassname()) .newInstance(); TilesUtil.setTilesUtil(impl); } catch (ClassCastException ex) { throw new ServletException( STR + getTilesUtilImplClassname() + STR + TilesUtilStrutsImpl.class.getName() + "'", ex); } catch (Exception ex) { throw new ServletException( STR, ex); } } }
|
/**
* Set TilesUtil implementation according to properties 'tilesUtilImplClassname'
* and 'moduleAware'. These properties are taken into account only once. A
* side effect is that only the values set in the first initialized plugin are
* effectively taken into account.
* @throws ServletException
*/
|
Set TilesUtil implementation according to properties 'tilesUtilImplClassname' and 'moduleAware'. These properties are taken into account only once. A side effect is that only the values set in the first initialized plugin are effectively taken into account
|
initTilesUtil
|
{
"repo_name": "davcamer/clients",
"path": "projects-for-testing/struts/tiles/src/main/java/org/apache/struts/tiles/TilesPlugin.java",
"license": "apache-2.0",
"size": 15064
}
|
[
"javax.servlet.ServletException",
"org.apache.struts.util.RequestUtils"
] |
import javax.servlet.ServletException; import org.apache.struts.util.RequestUtils;
|
import javax.servlet.*; import org.apache.struts.util.*;
|
[
"javax.servlet",
"org.apache.struts"
] |
javax.servlet; org.apache.struts;
| 1,271,633
|
public FXMatrix getFXMatrix() {
return _fxMatrix;
}
|
FXMatrix function() { return _fxMatrix; }
|
/**
* Gets the fxMatrix field.
* @return the fxMatrix
*/
|
Gets the fxMatrix field
|
getFXMatrix
|
{
"repo_name": "McLeodMoores/starling",
"path": "projects/analytics/src/main/java/com/opengamma/analytics/financial/provider/method/SuccessiveRootFinderCalibrationObjective.java",
"license": "apache-2.0",
"size": 4621
}
|
[
"com.opengamma.analytics.financial.forex.method.FXMatrix"
] |
import com.opengamma.analytics.financial.forex.method.FXMatrix;
|
import com.opengamma.analytics.financial.forex.method.*;
|
[
"com.opengamma.analytics"
] |
com.opengamma.analytics;
| 92,936
|
public static XContentBuilder contentBuilder(XContentType type) throws IOException {
if (type == XContentType.JSON) {
return JsonXContent.contentBuilder();
} else if (type == XContentType.SMILE) {
return SmileXContent.contentBuilder();
} else if (type == XContentType.YAML) {
return YamlXContent.contentBuilder();
} else if (type == XContentType.CBOR) {
return CborXContent.contentBuilder();
}
throw new IllegalArgumentException("No matching content type for " + type);
}
|
static XContentBuilder function(XContentType type) throws IOException { if (type == XContentType.JSON) { return JsonXContent.contentBuilder(); } else if (type == XContentType.SMILE) { return SmileXContent.contentBuilder(); } else if (type == XContentType.YAML) { return YamlXContent.contentBuilder(); } else if (type == XContentType.CBOR) { return CborXContent.contentBuilder(); } throw new IllegalArgumentException(STR + type); }
|
/**
* Returns a binary content builder for the provided content type.
*/
|
Returns a binary content builder for the provided content type
|
contentBuilder
|
{
"repo_name": "weipinghe/elasticsearch",
"path": "core/src/main/java/org/elasticsearch/common/xcontent/XContentFactory.java",
"license": "apache-2.0",
"size": 12991
}
|
[
"java.io.IOException",
"org.elasticsearch.common.xcontent.cbor.CborXContent",
"org.elasticsearch.common.xcontent.json.JsonXContent",
"org.elasticsearch.common.xcontent.smile.SmileXContent",
"org.elasticsearch.common.xcontent.yaml.YamlXContent"
] |
import java.io.IOException; import org.elasticsearch.common.xcontent.cbor.CborXContent; import org.elasticsearch.common.xcontent.json.JsonXContent; import org.elasticsearch.common.xcontent.smile.SmileXContent; import org.elasticsearch.common.xcontent.yaml.YamlXContent;
|
import java.io.*; import org.elasticsearch.common.xcontent.cbor.*; import org.elasticsearch.common.xcontent.json.*; import org.elasticsearch.common.xcontent.smile.*; import org.elasticsearch.common.xcontent.yaml.*;
|
[
"java.io",
"org.elasticsearch.common"
] |
java.io; org.elasticsearch.common;
| 247,498
|
Property createProperty(String name, Reader value) throws SearchEngineException;
|
Property createProperty(String name, Reader value) throws SearchEngineException;
|
/**
* Creates a property (TEXT type) for the specified reader.
*/
|
Creates a property (TEXT type) for the specified reader
|
createProperty
|
{
"repo_name": "unkascrack/compass-fork",
"path": "compass-core/src/main/java/org/compass/core/ResourceFactory.java",
"license": "apache-2.0",
"size": 3818
}
|
[
"java.io.Reader",
"org.compass.core.engine.SearchEngineException"
] |
import java.io.Reader; import org.compass.core.engine.SearchEngineException;
|
import java.io.*; import org.compass.core.engine.*;
|
[
"java.io",
"org.compass.core"
] |
java.io; org.compass.core;
| 2,134,285
|
public static String[] getNames(JSONObject jo) {
int length = jo.length();
if (length == 0) {
return null;
}
Iterator iterator = jo.keys();
String[] names = new String[length];
int i = 0;
while (iterator.hasNext()) {
names[i] = (String) iterator.next();
i += 1;
}
return names;
}
|
static String[] function(JSONObject jo) { int length = jo.length(); if (length == 0) { return null; } Iterator iterator = jo.keys(); String[] names = new String[length]; int i = 0; while (iterator.hasNext()) { names[i] = (String) iterator.next(); i += 1; } return names; }
|
/**
* Get an array of field names from a JSONObject.
*
* @return An array of field names, or null if there are no names.
*/
|
Get an array of field names from a JSONObject
|
getNames
|
{
"repo_name": "FreeSchoolHackers/RiTa",
"path": "java/rita/json/JSONObject.java",
"license": "gpl-3.0",
"size": 56711
}
|
[
"java.util.Iterator"
] |
import java.util.Iterator;
|
import java.util.*;
|
[
"java.util"
] |
java.util;
| 1,423,593
|
@Override
public KeyParameter deriveKey(CharSequence password) throws KeyCrypterException {
byte[] passwordBytes = null;
try {
passwordBytes = convertToByteArray(password);
byte[] salt = new byte[0];
if ( scryptParameters.getSalt() != null) {
salt = scryptParameters.getSalt().toByteArray();
} else {
// Warn the user that they are not using a salt.
// (Some early MultiBit wallets had a blank salt).
log.warn("You are using a ScryptParameters with no salt. Your encryption may be vulnerable to a dictionary attack.");
}
final Stopwatch watch = Stopwatch.createStarted();
byte[] keyBytes = SCrypt.scrypt(passwordBytes, salt, (int) scryptParameters.getN(), scryptParameters.getR(), scryptParameters.getP(), KEY_LENGTH);
watch.stop();
log.info("Deriving key took {} for {} scrypt iterations.", watch, scryptParameters.getN());
return new KeyParameter(keyBytes);
} catch (Exception e) {
throw new KeyCrypterException("Could not generate key from password and salt.", e);
} finally {
// Zero the password bytes.
if (passwordBytes != null) {
java.util.Arrays.fill(passwordBytes, (byte) 0);
}
}
}
|
KeyParameter function(CharSequence password) throws KeyCrypterException { byte[] passwordBytes = null; try { passwordBytes = convertToByteArray(password); byte[] salt = new byte[0]; if ( scryptParameters.getSalt() != null) { salt = scryptParameters.getSalt().toByteArray(); } else { log.warn(STR); } final Stopwatch watch = Stopwatch.createStarted(); byte[] keyBytes = SCrypt.scrypt(passwordBytes, salt, (int) scryptParameters.getN(), scryptParameters.getR(), scryptParameters.getP(), KEY_LENGTH); watch.stop(); log.info(STR, watch, scryptParameters.getN()); return new KeyParameter(keyBytes); } catch (Exception e) { throw new KeyCrypterException(STR, e); } finally { if (passwordBytes != null) { java.util.Arrays.fill(passwordBytes, (byte) 0); } } }
|
/**
* Generate AES key.
*
* This is a very slow operation compared to encrypt/ decrypt so it is normally worth caching the result.
*
* @param password The password to use in key generation
* @return The KeyParameter containing the created AES key
* @throws KeyCrypterException
*/
|
Generate AES key. This is a very slow operation compared to encrypt/ decrypt so it is normally worth caching the result
|
deriveKey
|
{
"repo_name": "kmels/bitcoinj",
"path": "core/src/main/java/org/bitcoinj/crypto/KeyCrypterScrypt.java",
"license": "apache-2.0",
"size": 11123
}
|
[
"com.google.common.base.Stopwatch",
"com.lambdaworks.crypto.SCrypt",
"java.util.Arrays",
"org.bouncycastle.crypto.params.KeyParameter"
] |
import com.google.common.base.Stopwatch; import com.lambdaworks.crypto.SCrypt; import java.util.Arrays; import org.bouncycastle.crypto.params.KeyParameter;
|
import com.google.common.base.*; import com.lambdaworks.crypto.*; import java.util.*; import org.bouncycastle.crypto.params.*;
|
[
"com.google.common",
"com.lambdaworks.crypto",
"java.util",
"org.bouncycastle.crypto"
] |
com.google.common; com.lambdaworks.crypto; java.util; org.bouncycastle.crypto;
| 403,726
|
public static void logSuccess(String user, String operation, String target,
ApplicationId appId, ContainerId containerId) {
if (LOG.isInfoEnabled()) {
LOG.info(createSuccessLog(user, operation, target, appId, null,
containerId));
}
}
|
static void function(String user, String operation, String target, ApplicationId appId, ContainerId containerId) { if (LOG.isInfoEnabled()) { LOG.info(createSuccessLog(user, operation, target, appId, null, containerId)); } }
|
/**
* Create a readable and parseable audit log string for a successful event.
*
* @param user User who made the service request to the ResourceManager
* @param operation Operation requested by the user.
* @param target The target on which the operation is being performed.
* @param appId Application Id in which operation was performed.
* @param containerId Container Id in which operation was performed.
*
* <br><br>
* Note that the {@link RMAuditLogger} uses tabs ('\t') as a key-val delimiter
* and hence the value fields should not contains tabs ('\t').
*/
|
Create a readable and parseable audit log string for a successful event
|
logSuccess
|
{
"repo_name": "laxman-ch/hadoop",
"path": "hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/RMAuditLogger.java",
"license": "apache-2.0",
"size": 12567
}
|
[
"org.apache.hadoop.yarn.api.records.ApplicationId",
"org.apache.hadoop.yarn.api.records.ContainerId"
] |
import org.apache.hadoop.yarn.api.records.ApplicationId; import org.apache.hadoop.yarn.api.records.ContainerId;
|
import org.apache.hadoop.yarn.api.records.*;
|
[
"org.apache.hadoop"
] |
org.apache.hadoop;
| 335,870
|
public Object getValueAt(int rowIndex, int columnIndex) {
if (null == sanpv) {
sanpv = project.getSaNpvTable();
}
if (0 >= columnIndex) {
Mmf mmf = project.get(rowIndex);
return mmf.getId() + ": " + mmf.getName();
} else {
return sanpv[rowIndex][columnIndex - 1];
}
}
|
Object function(int rowIndex, int columnIndex) { if (null == sanpv) { sanpv = project.getSaNpvTable(); } if (0 >= columnIndex) { Mmf mmf = project.get(rowIndex); return mmf.getId() + STR + mmf.getName(); } else { return sanpv[rowIndex][columnIndex - 1]; } }
|
/**
* Method for getting the SaNpv value for a mmf at a period
*
* If column is 0 returns id and name for the mmf.
*
* @see no.ntnu.mmfplanner.model.Project#getSaNpvTable()
* @param rowIndex the index of the row (mmf)
* @param columnIndex the index of the column (period)
*/
|
Method for getting the SaNpv value for a mmf at a period If column is 0 returns id and name for the mmf
|
getValueAt
|
{
"repo_name": "jodal/mmfplanner",
"path": "src/main/java/no/ntnu/mmfplanner/ui/model/SaNpvTableModel.java",
"license": "gpl-2.0",
"size": 3150
}
|
[
"no.ntnu.mmfplanner.model.Mmf"
] |
import no.ntnu.mmfplanner.model.Mmf;
|
import no.ntnu.mmfplanner.model.*;
|
[
"no.ntnu.mmfplanner"
] |
no.ntnu.mmfplanner;
| 342,917
|
@ServiceMethod(returns = ReturnType.SINGLE)
private Mono<Response<Flux<ByteBuffer>>> updateSegmentsWithResponseAsync(
String resourceGroupName,
String privateCloudName,
String segmentId,
WorkloadNetworkSegmentInner workloadNetworkSegment,
Context context) {
if (this.client.getEndpoint() == null) {
return Mono
.error(
new IllegalArgumentException(
"Parameter this.client.getEndpoint() is required and cannot be null."));
}
if (this.client.getSubscriptionId() == null) {
return Mono
.error(
new IllegalArgumentException(
"Parameter this.client.getSubscriptionId() is required and cannot be null."));
}
if (resourceGroupName == null) {
return Mono
.error(new IllegalArgumentException("Parameter resourceGroupName is required and cannot be null."));
}
if (privateCloudName == null) {
return Mono
.error(new IllegalArgumentException("Parameter privateCloudName is required and cannot be null."));
}
if (segmentId == null) {
return Mono.error(new IllegalArgumentException("Parameter segmentId is required and cannot be null."));
}
if (workloadNetworkSegment == null) {
return Mono
.error(
new IllegalArgumentException("Parameter workloadNetworkSegment is required and cannot be null."));
} else {
workloadNetworkSegment.validate();
}
final String accept = "application/json";
context = this.client.mergeContext(context);
return service
.updateSegments(
this.client.getEndpoint(),
this.client.getSubscriptionId(),
resourceGroupName,
this.client.getApiVersion(),
privateCloudName,
segmentId,
workloadNetworkSegment,
accept,
context);
}
|
@ServiceMethod(returns = ReturnType.SINGLE) Mono<Response<Flux<ByteBuffer>>> function( String resourceGroupName, String privateCloudName, String segmentId, WorkloadNetworkSegmentInner workloadNetworkSegment, Context context) { if (this.client.getEndpoint() == null) { return Mono .error( new IllegalArgumentException( STR)); } if (this.client.getSubscriptionId() == null) { return Mono .error( new IllegalArgumentException( STR)); } if (resourceGroupName == null) { return Mono .error(new IllegalArgumentException(STR)); } if (privateCloudName == null) { return Mono .error(new IllegalArgumentException(STR)); } if (segmentId == null) { return Mono.error(new IllegalArgumentException(STR)); } if (workloadNetworkSegment == null) { return Mono .error( new IllegalArgumentException(STR)); } else { workloadNetworkSegment.validate(); } final String accept = STR; context = this.client.mergeContext(context); return service .updateSegments( this.client.getEndpoint(), this.client.getSubscriptionId(), resourceGroupName, this.client.getApiVersion(), privateCloudName, segmentId, workloadNetworkSegment, accept, context); }
|
/**
* Create or update a segment by id in a private cloud workload network.
*
* @param resourceGroupName The name of the resource group. The name is case insensitive.
* @param privateCloudName Name of the private cloud.
* @param segmentId NSX Segment identifier. Generally the same as the Segment's display name.
* @param workloadNetworkSegment NSX Segment.
* @param context The context to associate with this operation.
* @throws IllegalArgumentException thrown if parameters fail the validation.
* @throws ManagementException thrown if the request is rejected by server.
* @throws RuntimeException all other wrapped checked exceptions if the request fails to be sent.
* @return nSX Segment.
*/
|
Create or update a segment by id in a private cloud workload network
|
updateSegmentsWithResponseAsync
|
{
"repo_name": "Azure/azure-sdk-for-java",
"path": "sdk/avs/azure-resourcemanager-avs/src/main/java/com/azure/resourcemanager/avs/implementation/WorkloadNetworksClientImpl.java",
"license": "mit",
"size": 538828
}
|
[
"com.azure.core.annotation.ReturnType",
"com.azure.core.annotation.ServiceMethod",
"com.azure.core.http.rest.Response",
"com.azure.core.util.Context",
"com.azure.resourcemanager.avs.fluent.models.WorkloadNetworkSegmentInner",
"java.nio.ByteBuffer"
] |
import com.azure.core.annotation.ReturnType; import com.azure.core.annotation.ServiceMethod; import com.azure.core.http.rest.Response; import com.azure.core.util.Context; import com.azure.resourcemanager.avs.fluent.models.WorkloadNetworkSegmentInner; import java.nio.ByteBuffer;
|
import com.azure.core.annotation.*; import com.azure.core.http.rest.*; import com.azure.core.util.*; import com.azure.resourcemanager.avs.fluent.models.*; import java.nio.*;
|
[
"com.azure.core",
"com.azure.resourcemanager",
"java.nio"
] |
com.azure.core; com.azure.resourcemanager; java.nio;
| 1,416,338
|
@Test
public void testTC11() throws Exception {
final Path p = new Path("/TC11/foo");
System.out.println("p=" + p);
//a. Create file and write one block of data. Close file.
final int len1 = (int)BLOCK_SIZE;
{
FSDataOutputStream out = fs.create(p, false, buffersize, REPLICATION, BLOCK_SIZE);
AppendTestUtil.write(out, 0, len1);
out.close();
}
//b. Reopen file in "append" mode. Append half block of data.
FSDataOutputStream out = fs.append(p);
final int len2 = (int)BLOCK_SIZE/2;
AppendTestUtil.write(out, len1, len2);
out.hflush();
//c. Rename file to file.new.
final Path pnew = new Path(p + ".new");
assertTrue(fs.rename(p, pnew));
//d. Close file handle that was opened in (b).
out.close();
//check block sizes
final long len = fs.getFileStatus(pnew).getLen();
final LocatedBlocks locatedblocks = fs.dfs.getNamenode().getBlockLocations(pnew.toString(), 0L, len);
final int numblock = locatedblocks.locatedBlockCount();
for(int i = 0; i < numblock; i++) {
final LocatedBlock lb = locatedblocks.get(i);
final ExtendedBlock blk = lb.getBlock();
final long size = lb.getBlockSize();
if (i < numblock - 1) {
assertEquals(BLOCK_SIZE, size);
}
for(DatanodeInfo datanodeinfo : lb.getLocations()) {
final DataNode dn = cluster.getDataNode(datanodeinfo.getIpcPort());
final Block metainfo = DataNodeTestUtils.getFSDataset(dn).getStoredBlock(
blk.getBlockPoolId(), blk.getBlockId());
assertEquals(size, metainfo.getNumBytes());
}
}
}
|
void function() throws Exception { final Path p = new Path(STR); System.out.println("p=" + p); final int len1 = (int)BLOCK_SIZE; { FSDataOutputStream out = fs.create(p, false, buffersize, REPLICATION, BLOCK_SIZE); AppendTestUtil.write(out, 0, len1); out.close(); } FSDataOutputStream out = fs.append(p); final int len2 = (int)BLOCK_SIZE/2; AppendTestUtil.write(out, len1, len2); out.hflush(); final Path pnew = new Path(p + ".new"); assertTrue(fs.rename(p, pnew)); out.close(); final long len = fs.getFileStatus(pnew).getLen(); final LocatedBlocks locatedblocks = fs.dfs.getNamenode().getBlockLocations(pnew.toString(), 0L, len); final int numblock = locatedblocks.locatedBlockCount(); for(int i = 0; i < numblock; i++) { final LocatedBlock lb = locatedblocks.get(i); final ExtendedBlock blk = lb.getBlock(); final long size = lb.getBlockSize(); if (i < numblock - 1) { assertEquals(BLOCK_SIZE, size); } for(DatanodeInfo datanodeinfo : lb.getLocations()) { final DataNode dn = cluster.getDataNode(datanodeinfo.getIpcPort()); final Block metainfo = DataNodeTestUtils.getFSDataset(dn).getStoredBlock( blk.getBlockPoolId(), blk.getBlockId()); assertEquals(size, metainfo.getNumBytes()); } } }
|
/**
* TC11: Racing rename
* @throws IOException an exception might be thrown
*/
|
TC11: Racing rename
|
testTC11
|
{
"repo_name": "ZhangXFeng/hadoop",
"path": "src/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestFileAppend3.java",
"license": "apache-2.0",
"size": 14601
}
|
[
"org.apache.hadoop.fs.FSDataOutputStream",
"org.apache.hadoop.fs.Path",
"org.apache.hadoop.hdfs.protocol.Block",
"org.apache.hadoop.hdfs.protocol.DatanodeInfo",
"org.apache.hadoop.hdfs.protocol.ExtendedBlock",
"org.apache.hadoop.hdfs.protocol.LocatedBlock",
"org.apache.hadoop.hdfs.protocol.LocatedBlocks",
"org.apache.hadoop.hdfs.server.datanode.DataNode",
"org.apache.hadoop.hdfs.server.datanode.DataNodeTestUtils",
"org.junit.Assert"
] |
import org.apache.hadoop.fs.FSDataOutputStream; import org.apache.hadoop.fs.Path; import org.apache.hadoop.hdfs.protocol.Block; import org.apache.hadoop.hdfs.protocol.DatanodeInfo; import org.apache.hadoop.hdfs.protocol.ExtendedBlock; import org.apache.hadoop.hdfs.protocol.LocatedBlock; import org.apache.hadoop.hdfs.protocol.LocatedBlocks; import org.apache.hadoop.hdfs.server.datanode.DataNode; import org.apache.hadoop.hdfs.server.datanode.DataNodeTestUtils; import org.junit.Assert;
|
import org.apache.hadoop.fs.*; import org.apache.hadoop.hdfs.protocol.*; import org.apache.hadoop.hdfs.server.datanode.*; import org.junit.*;
|
[
"org.apache.hadoop",
"org.junit"
] |
org.apache.hadoop; org.junit;
| 596,163
|
public static java.util.List extractCleansingandDressingComponentList(ims.domain.ILightweightDomainFactory domainFactory, ims.spinalinjuries.vo.NurAssessmentCleansingDressingVoCollection voCollection)
{
return extractCleansingandDressingComponentList(domainFactory, voCollection, null, new HashMap());
}
|
static java.util.List function(ims.domain.ILightweightDomainFactory domainFactory, ims.spinalinjuries.vo.NurAssessmentCleansingDressingVoCollection voCollection) { return extractCleansingandDressingComponentList(domainFactory, voCollection, null, new HashMap()); }
|
/**
* Create the ims.nursing.assessment.domain.objects.CleansingandDressingComponent list from the value object collection.
* @param domainFactory - used to create existing (persistent) domain objects.
* @param voCollection - the collection of value objects
*/
|
Create the ims.nursing.assessment.domain.objects.CleansingandDressingComponent list from the value object collection
|
extractCleansingandDressingComponentList
|
{
"repo_name": "FreudianNM/openMAXIMS",
"path": "Source Library/openmaxims_workspace/ValueObjects/src/ims/spinalinjuries/vo/domain/NurAssessmentCleansingDressingVoAssembler.java",
"license": "agpl-3.0",
"size": 25021
}
|
[
"java.util.HashMap"
] |
import java.util.HashMap;
|
import java.util.*;
|
[
"java.util"
] |
java.util;
| 622,678
|
public static void getBrokerSchemaFolderName(JET2Context context, String select, String variable) {
XPathContextExtender extender = XPathContextExtender.getInstance(context);
Object currentContext = extender.currentXPathContextObject();
Object element = extender.resolveSingle(currentContext, select);
String brokerSchema = extender.getContent(element);
if (brokerSchema != null) {
brokerSchema = brokerSchema.replace(".", "/");
}
context.setVariable(variable, brokerSchema);
}
|
static void function(JET2Context context, String select, String variable) { XPathContextExtender extender = XPathContextExtender.getInstance(context); Object currentContext = extender.currentXPathContextObject(); Object element = extender.resolveSingle(currentContext, select); String brokerSchema = extender.getContent(element); if (brokerSchema != null) { brokerSchema = brokerSchema.replace(".", "/"); } context.setVariable(variable, brokerSchema); }
|
/**
* Transform broker schema to schema folder name
*
* @param context
* <code>JET2Context</code>.
* @param brokerSchema
* The broker schema.
* @param variable
* Where to put the resulting value.
*/
|
Transform broker schema to schema folder name
|
getBrokerSchemaFolderName
|
{
"repo_name": "ot4i/service-facade-mq-request-response-pattern",
"path": "src/com.ibm.etools.mft.pattern.sen/src/com/ibm/etools/mft/pattern/sen/plugin/PatternUtility.java",
"license": "epl-1.0",
"size": 9252
}
|
[
"org.eclipse.jet.JET2Context",
"org.eclipse.jet.XPathContextExtender"
] |
import org.eclipse.jet.JET2Context; import org.eclipse.jet.XPathContextExtender;
|
import org.eclipse.jet.*;
|
[
"org.eclipse.jet"
] |
org.eclipse.jet;
| 2,681,424
|
public Choice getChoice() {
return new Choice(this);
}
public List<Unavailability> getUnavailabilities() { return iUnavailabilities; }
|
Choice function() { return new Choice(this); } public List<Unavailability> getUnavailabilities() { return iUnavailabilities; }
|
/**
* Choice matching this section
* @return choice matching this section
*/
|
Choice matching this section
|
getChoice
|
{
"repo_name": "UniTime/cpsolver",
"path": "src/org/cpsolver/studentsct/model/Section.java",
"license": "lgpl-3.0",
"size": 38061
}
|
[
"java.util.List"
] |
import java.util.List;
|
import java.util.*;
|
[
"java.util"
] |
java.util;
| 1,288,542
|
private JPanel getJStrainColorsPanel() {
if (jStrainColorsPanel == null) {
jStrainColorsPanel = new JPanel();
SpringLayout layout = new SpringLayout();
jStrainColorsPanel.setLayout(layout);
jStrainColorsPanel.setBorder(BorderFactory.createRaisedBevelBorder());
jStrainColorsPanel.add(getJStrainConstantCheckPanel(), null);
jStrainColorsPanel.add(getJStrainConstantSettingsPanel(), null);
jStrainColorsPanel.add(getJStrainRandomCheckPanel(), null);
jStrainColorsPanel.add(getJStrainRandomSettingsPanel(), null);
jStrainColorsPanel.add(getJStrainTintCheckPanel(), null);
jStrainColorsPanel.add(getJStrainTintSettingsPanel(), null);
jStrainColorsPanel.add(getJStrainTwoToneCheckPanel(), null);
jStrainColorsPanel.add(getJStrainTwoToneSettingsPanel(), null);
Util.makeCompactGrid(jStrainColorsPanel,layout,4,2,5,5,5,5);
}
return jStrainColorsPanel;
}
|
JPanel function() { if (jStrainColorsPanel == null) { jStrainColorsPanel = new JPanel(); SpringLayout layout = new SpringLayout(); jStrainColorsPanel.setLayout(layout); jStrainColorsPanel.setBorder(BorderFactory.createRaisedBevelBorder()); jStrainColorsPanel.add(getJStrainConstantCheckPanel(), null); jStrainColorsPanel.add(getJStrainConstantSettingsPanel(), null); jStrainColorsPanel.add(getJStrainRandomCheckPanel(), null); jStrainColorsPanel.add(getJStrainRandomSettingsPanel(), null); jStrainColorsPanel.add(getJStrainTintCheckPanel(), null); jStrainColorsPanel.add(getJStrainTintSettingsPanel(), null); jStrainColorsPanel.add(getJStrainTwoToneCheckPanel(), null); jStrainColorsPanel.add(getJStrainTwoToneSettingsPanel(), null); Util.makeCompactGrid(jStrainColorsPanel,layout,4,2,5,5,5,5); } return jStrainColorsPanel; }
|
/**
* This method initializes jStrainColorsPanel
*
* @return javax.swing.JPanel
*/
|
This method initializes jStrainColorsPanel
|
getJStrainColorsPanel
|
{
"repo_name": "jmeppley/strainer",
"path": "src/amd/strainer/display/actions/DisplayOptionsDialog.java",
"license": "lgpl-3.0",
"size": 56827
}
|
[
"javax.swing.BorderFactory",
"javax.swing.JPanel",
"javax.swing.SpringLayout"
] |
import javax.swing.BorderFactory; import javax.swing.JPanel; import javax.swing.SpringLayout;
|
import javax.swing.*;
|
[
"javax.swing"
] |
javax.swing;
| 1,987,878
|
public PointF hDGetPoint(Q Dcon) {
return (hDGetHex(Dcon).getLoc());
};
|
PointF function(Q Dcon) { return (hDGetHex(Dcon).getLoc()); };
|
/**
* Gets the point associated with the HD port.
*/
|
Gets the point associated with the HD port
|
hDGetPoint
|
{
"repo_name": "viridian1138/VectorVictor",
"path": "VectorVictor/VectorVictor/src/geomdir/depictors/Vect2Base.java",
"license": "gpl-3.0",
"size": 61111
}
|
[
"android.graphics.PointF"
] |
import android.graphics.PointF;
|
import android.graphics.*;
|
[
"android.graphics"
] |
android.graphics;
| 1,474,878
|
public static final Parcelable.Creator<CharSequence> CHAR_SEQUENCE_CREATOR
= new Parcelable.Creator<CharSequence>() {
public CharSequence createFromParcel(Parcel p) {
int kind = p.readInt();
String string = p.readString();
if (string == null) {
return null;
}
if (kind == 1) {
return string;
}
SpannableString sp = new SpannableString(string);
while (true) {
kind = p.readInt();
if (kind == 0)
break;
switch (kind) {
case ALIGNMENT_SPAN:
readSpan(p, sp, new AlignmentSpan.Standard(p));
break;
case FOREGROUND_COLOR_SPAN:
readSpan(p, sp, new ForegroundColorSpan(p));
break;
case RELATIVE_SIZE_SPAN:
readSpan(p, sp, new RelativeSizeSpan(p));
break;
case SCALE_X_SPAN:
readSpan(p, sp, new ScaleXSpan(p));
break;
case STRIKETHROUGH_SPAN:
readSpan(p, sp, new StrikethroughSpan(p));
break;
case UNDERLINE_SPAN:
readSpan(p, sp, new UnderlineSpan(p));
break;
case STYLE_SPAN:
readSpan(p, sp, new StyleSpan(p));
break;
case BULLET_SPAN:
readSpan(p, sp, new BulletSpan(p));
break;
case QUOTE_SPAN:
readSpan(p, sp, new QuoteSpan(p));
break;
case LEADING_MARGIN_SPAN:
readSpan(p, sp, new LeadingMarginSpan.Standard(p));
break;
case URL_SPAN:
readSpan(p, sp, new URLSpan(p));
break;
case BACKGROUND_COLOR_SPAN:
readSpan(p, sp, new BackgroundColorSpan(p));
break;
case TYPEFACE_SPAN:
readSpan(p, sp, new TypefaceSpan(p));
break;
case SUPERSCRIPT_SPAN:
readSpan(p, sp, new SuperscriptSpan(p));
break;
case SUBSCRIPT_SPAN:
readSpan(p, sp, new SubscriptSpan(p));
break;
case ABSOLUTE_SIZE_SPAN:
readSpan(p, sp, new AbsoluteSizeSpan(p));
break;
case TEXT_APPEARANCE_SPAN:
readSpan(p, sp, new TextAppearanceSpan(p));
break;
case ANNOTATION:
readSpan(p, sp, new Annotation(p));
break;
case SUGGESTION_SPAN:
readSpan(p, sp, new SuggestionSpan(p));
break;
case SPELL_CHECK_SPAN:
readSpan(p, sp, new SpellCheckSpan(p));
break;
case SUGGESTION_RANGE_SPAN:
readSpan(p, sp, new SuggestionRangeSpan(p));
break;
case EASY_EDIT_SPAN:
readSpan(p, sp, new EasyEditSpan(p));
break;
case LOCALE_SPAN:
readSpan(p, sp, new LocaleSpan(p));
break;
default:
throw new RuntimeException("bogus span encoding " + kind);
}
}
return sp;
}
|
static final Parcelable.Creator<CharSequence> CHAR_SEQUENCE_CREATOR = new Parcelable.Creator<CharSequence>() { public CharSequence function(Parcel p) { int kind = p.readInt(); String string = p.readString(); if (string == null) { return null; } if (kind == 1) { return string; } SpannableString sp = new SpannableString(string); while (true) { kind = p.readInt(); if (kind == 0) break; switch (kind) { case ALIGNMENT_SPAN: readSpan(p, sp, new AlignmentSpan.Standard(p)); break; case FOREGROUND_COLOR_SPAN: readSpan(p, sp, new ForegroundColorSpan(p)); break; case RELATIVE_SIZE_SPAN: readSpan(p, sp, new RelativeSizeSpan(p)); break; case SCALE_X_SPAN: readSpan(p, sp, new ScaleXSpan(p)); break; case STRIKETHROUGH_SPAN: readSpan(p, sp, new StrikethroughSpan(p)); break; case UNDERLINE_SPAN: readSpan(p, sp, new UnderlineSpan(p)); break; case STYLE_SPAN: readSpan(p, sp, new StyleSpan(p)); break; case BULLET_SPAN: readSpan(p, sp, new BulletSpan(p)); break; case QUOTE_SPAN: readSpan(p, sp, new QuoteSpan(p)); break; case LEADING_MARGIN_SPAN: readSpan(p, sp, new LeadingMarginSpan.Standard(p)); break; case URL_SPAN: readSpan(p, sp, new URLSpan(p)); break; case BACKGROUND_COLOR_SPAN: readSpan(p, sp, new BackgroundColorSpan(p)); break; case TYPEFACE_SPAN: readSpan(p, sp, new TypefaceSpan(p)); break; case SUPERSCRIPT_SPAN: readSpan(p, sp, new SuperscriptSpan(p)); break; case SUBSCRIPT_SPAN: readSpan(p, sp, new SubscriptSpan(p)); break; case ABSOLUTE_SIZE_SPAN: readSpan(p, sp, new AbsoluteSizeSpan(p)); break; case TEXT_APPEARANCE_SPAN: readSpan(p, sp, new TextAppearanceSpan(p)); break; case ANNOTATION: readSpan(p, sp, new Annotation(p)); break; case SUGGESTION_SPAN: readSpan(p, sp, new SuggestionSpan(p)); break; case SPELL_CHECK_SPAN: readSpan(p, sp, new SpellCheckSpan(p)); break; case SUGGESTION_RANGE_SPAN: readSpan(p, sp, new SuggestionRangeSpan(p)); break; case EASY_EDIT_SPAN: readSpan(p, sp, new EasyEditSpan(p)); break; case LOCALE_SPAN: readSpan(p, sp, new LocaleSpan(p)); break; default: throw new RuntimeException(STR + kind); } } return sp; }
|
/**
* Read and return a new CharSequence, possibly with styles,
* from the parcel.
*/
|
Read and return a new CharSequence, possibly with styles, from the parcel
|
createFromParcel
|
{
"repo_name": "JSDemos/android-sdk-20",
"path": "src/android/text/TextUtils.java",
"license": "apache-2.0",
"size": 59445
}
|
[
"android.os.Parcel",
"android.os.Parcelable",
"android.text.style.AbsoluteSizeSpan",
"android.text.style.AlignmentSpan",
"android.text.style.BackgroundColorSpan",
"android.text.style.BulletSpan",
"android.text.style.EasyEditSpan",
"android.text.style.ForegroundColorSpan",
"android.text.style.LeadingMarginSpan",
"android.text.style.LocaleSpan",
"android.text.style.QuoteSpan",
"android.text.style.RelativeSizeSpan",
"android.text.style.ScaleXSpan",
"android.text.style.SpellCheckSpan",
"android.text.style.StrikethroughSpan",
"android.text.style.StyleSpan",
"android.text.style.SubscriptSpan",
"android.text.style.SuggestionRangeSpan",
"android.text.style.SuggestionSpan",
"android.text.style.SuperscriptSpan",
"android.text.style.TextAppearanceSpan",
"android.text.style.TypefaceSpan",
"android.text.style.URLSpan",
"android.text.style.UnderlineSpan"
] |
import android.os.Parcel; import android.os.Parcelable; import android.text.style.AbsoluteSizeSpan; import android.text.style.AlignmentSpan; import android.text.style.BackgroundColorSpan; import android.text.style.BulletSpan; import android.text.style.EasyEditSpan; import android.text.style.ForegroundColorSpan; import android.text.style.LeadingMarginSpan; import android.text.style.LocaleSpan; import android.text.style.QuoteSpan; import android.text.style.RelativeSizeSpan; import android.text.style.ScaleXSpan; import android.text.style.SpellCheckSpan; import android.text.style.StrikethroughSpan; import android.text.style.StyleSpan; import android.text.style.SubscriptSpan; import android.text.style.SuggestionRangeSpan; import android.text.style.SuggestionSpan; import android.text.style.SuperscriptSpan; import android.text.style.TextAppearanceSpan; import android.text.style.TypefaceSpan; import android.text.style.URLSpan; import android.text.style.UnderlineSpan;
|
import android.os.*; import android.text.style.*;
|
[
"android.os",
"android.text"
] |
android.os; android.text;
| 2,654,936
|
public void apply( ICurveWithNormals3f sweep , Point2f data[] , float lateralOffset , float angle , Point3f trans[] )
{
if( data.length != trans.length )
{
throw new IllegalArgumentException( "result must be the same size as data" );
}
for( int i = 0 ; i < data.length ; i++ )
{
apply( sweep , data[ i ].x , data[ i ].y + lateralOffset , angle , trans[ i ] );
}
}
|
void function( ICurveWithNormals3f sweep , Point2f data[] , float lateralOffset , float angle , Point3f trans[] ) { if( data.length != trans.length ) { throw new IllegalArgumentException( STR ); } for( int i = 0 ; i < data.length ; i++ ) { apply( sweep , data[ i ].x , data[ i ].y + lateralOffset , angle , trans[ i ] ); } }
|
/**
* Transforms a collection of depth/offset points on the sweep into 3D world coordinates, storing them in the result collection. The result must be the same
* size as the data and contain no null elements.
*
* @return whether the operation was successful or not.
*/
|
Transforms a collection of depth/offset points on the sweep into 3D world coordinates, storing them in the result collection. The result must be the same size as the data and contain no null elements
|
apply
|
{
"repo_name": "jedwards1211/breakout",
"path": "andork-j3d-utils/src/org/andork/math3d/curve/Sweeper3f.java",
"license": "gpl-2.0",
"size": 8589
}
|
[
"javax.vecmath.Point2f",
"javax.vecmath.Point3f",
"org.andork.math.curve.ICurveWithNormals3f"
] |
import javax.vecmath.Point2f; import javax.vecmath.Point3f; import org.andork.math.curve.ICurveWithNormals3f;
|
import javax.vecmath.*; import org.andork.math.curve.*;
|
[
"javax.vecmath",
"org.andork.math"
] |
javax.vecmath; org.andork.math;
| 1,180,867
|
public TransLogTable getTransLogTable() {
return transLogTable;
}
|
TransLogTable function() { return transLogTable; }
|
/**
* Gets the log table for the transformation.
*
* @return the log table for the transformation
*/
|
Gets the log table for the transformation
|
getTransLogTable
|
{
"repo_name": "eayoungs/pentaho-kettle",
"path": "engine/src/org/pentaho/di/trans/TransMeta.java",
"license": "apache-2.0",
"size": 221441
}
|
[
"org.pentaho.di.core.logging.TransLogTable"
] |
import org.pentaho.di.core.logging.TransLogTable;
|
import org.pentaho.di.core.logging.*;
|
[
"org.pentaho.di"
] |
org.pentaho.di;
| 2,316,642
|
public boolean isIncludeParticipants() {
Set<ISharePreference> prefs = getPreferencesByType(IncludeParticipantsPreference.INCLUDE_PARTICIPANTS);
for(ISharePreference pref: prefs) {
return Boolean.parseBoolean(pref.getValue());
}
// preference not present, default is false
return false;
}
|
boolean function() { Set<ISharePreference> prefs = getPreferencesByType(IncludeParticipantsPreference.INCLUDE_PARTICIPANTS); for(ISharePreference pref: prefs) { return Boolean.parseBoolean(pref.getValue()); } return false; }
|
/**
* Short cut to determine if this share has an
* IncludeParticipants preference set to true.
*
* @return the value of the IncludeParticipants preference, or false if not set
*/
|
Short cut to determine if this share has an IncludeParticipants preference set to true
|
isIncludeParticipants
|
{
"repo_name": "nblair/shareurl",
"path": "src/main/java/edu/wisc/wisccal/shareurl/domain/SharePreferences.java",
"license": "apache-2.0",
"size": 14237
}
|
[
"java.util.Set"
] |
import java.util.Set;
|
import java.util.*;
|
[
"java.util"
] |
java.util;
| 320,469
|
// fill grid with background color
final int bgR = (int) (this.backgroundCol >> 16);
final int bgG = (int) ((this.backgroundCol >> 8) & 0xff);
final int bgB = (int) (this.backgroundCol & 0xff);
if (this.frame == null) {
final Graphics2D gr = this.image.createGraphics();
Color c = new Color(bgR, bgG, bgB);
gr.setBackground(c);
gr.clearRect(0, 0, this.width, this.height);
gr.setColor(c);
gr.fillRect(0, 0, this.width, this.height);
} else {
int p = 0;
for (int i = 0; i < width; i++) {
this.frame[p++] = (byte) bgR;
this.frame[p++] = (byte) bgG;
this.frame[p++] = (byte) bgB;
}
final int rw = width * 3;
for (int i = 1; i < height; i++) {
System.arraycopy(this.frame, 0, this.frame, i * rw, rw);
}
}
}
|
final int bgR = (int) (this.backgroundCol >> 16); final int bgG = (int) ((this.backgroundCol >> 8) & 0xff); final int bgB = (int) (this.backgroundCol & 0xff); if (this.frame == null) { final Graphics2D gr = this.image.createGraphics(); Color c = new Color(bgR, bgG, bgB); gr.setBackground(c); gr.clearRect(0, 0, this.width, this.height); gr.setColor(c); gr.fillRect(0, 0, this.width, this.height); } else { int p = 0; for (int i = 0; i < width; i++) { this.frame[p++] = (byte) bgR; this.frame[p++] = (byte) bgG; this.frame[p++] = (byte) bgB; } final int rw = width * 3; for (int i = 1; i < height; i++) { System.arraycopy(this.frame, 0, this.frame, i * rw, rw); } } }
|
/**
* Deletes all pixels of image and sets them to previously defined
* background color.
*/
|
Deletes all pixels of image and sets them to previously defined background color
|
clear
|
{
"repo_name": "karajrish/loklak_server",
"path": "src/org/loklak/visualization/graphics/RasterPlotter.java",
"license": "lgpl-2.1",
"size": 46617
}
|
[
"java.awt.Color",
"java.awt.Graphics2D"
] |
import java.awt.Color; import java.awt.Graphics2D;
|
import java.awt.*;
|
[
"java.awt"
] |
java.awt;
| 1,469,302
|
public Observable<ServiceResponse<NetworkInterfaceIPConfigurationInner>> getVirtualMachineScaleSetIpConfigurationWithServiceResponseAsync(String resourceGroupName, String virtualMachineScaleSetName, String virtualmachineIndex, String networkInterfaceName, String ipConfigurationName) {
if (resourceGroupName == null) {
throw new IllegalArgumentException("Parameter resourceGroupName is required and cannot be null.");
}
if (virtualMachineScaleSetName == null) {
throw new IllegalArgumentException("Parameter virtualMachineScaleSetName is required and cannot be null.");
}
if (virtualmachineIndex == null) {
throw new IllegalArgumentException("Parameter virtualmachineIndex is required and cannot be null.");
}
if (networkInterfaceName == null) {
throw new IllegalArgumentException("Parameter networkInterfaceName is required and cannot be null.");
}
if (ipConfigurationName == null) {
throw new IllegalArgumentException("Parameter ipConfigurationName is required and cannot be null.");
}
if (this.client.subscriptionId() == null) {
throw new IllegalArgumentException("Parameter this.client.subscriptionId() is required and cannot be null.");
}
|
Observable<ServiceResponse<NetworkInterfaceIPConfigurationInner>> function(String resourceGroupName, String virtualMachineScaleSetName, String virtualmachineIndex, String networkInterfaceName, String ipConfigurationName) { if (resourceGroupName == null) { throw new IllegalArgumentException(STR); } if (virtualMachineScaleSetName == null) { throw new IllegalArgumentException(STR); } if (virtualmachineIndex == null) { throw new IllegalArgumentException(STR); } if (networkInterfaceName == null) { throw new IllegalArgumentException(STR); } if (ipConfigurationName == null) { throw new IllegalArgumentException(STR); } if (this.client.subscriptionId() == null) { throw new IllegalArgumentException(STR); }
|
/**
* Get the specified network interface ip configuration in a virtual machine scale set.
*
* @param resourceGroupName The name of the resource group.
* @param virtualMachineScaleSetName The name of the virtual machine scale set.
* @param virtualmachineIndex The virtual machine index.
* @param networkInterfaceName The name of the network interface.
* @param ipConfigurationName The name of the ip configuration.
* @throws IllegalArgumentException thrown if parameters fail the validation
* @return the observable to the NetworkInterfaceIPConfigurationInner object
*/
|
Get the specified network interface ip configuration in a virtual machine scale set
|
getVirtualMachineScaleSetIpConfigurationWithServiceResponseAsync
|
{
"repo_name": "selvasingh/azure-sdk-for-java",
"path": "sdk/network/mgmt-v2019_08_01/src/main/java/com/microsoft/azure/management/network/v2019_08_01/implementation/NetworkInterfacesInner.java",
"license": "mit",
"size": 192401
}
|
[
"com.microsoft.rest.ServiceResponse"
] |
import com.microsoft.rest.ServiceResponse;
|
import com.microsoft.rest.*;
|
[
"com.microsoft.rest"
] |
com.microsoft.rest;
| 2,288,082
|
public RuleConfiguredTargetBuilder addOutputGroup(String name, NestedSet<Artifact> artifacts) {
getOutputGroupBuilder(name).addTransitive(artifacts);
return this;
}
|
RuleConfiguredTargetBuilder function(String name, NestedSet<Artifact> artifacts) { getOutputGroupBuilder(name).addTransitive(artifacts); return this; }
|
/**
* Adds a set of files to an output group.
*/
|
Adds a set of files to an output group
|
addOutputGroup
|
{
"repo_name": "damienmg/bazel",
"path": "src/main/java/com/google/devtools/build/lib/analysis/RuleConfiguredTargetBuilder.java",
"license": "apache-2.0",
"size": 16782
}
|
[
"com.google.devtools.build.lib.actions.Artifact",
"com.google.devtools.build.lib.collect.nestedset.NestedSet"
] |
import com.google.devtools.build.lib.actions.Artifact; import com.google.devtools.build.lib.collect.nestedset.NestedSet;
|
import com.google.devtools.build.lib.actions.*; import com.google.devtools.build.lib.collect.nestedset.*;
|
[
"com.google.devtools"
] |
com.google.devtools;
| 1,823,822
|
public Integer createReplyComment(ReplyCommentDTO replyCommentDTO);
|
Integer function(ReplyCommentDTO replyCommentDTO);
|
/**
* Creates the reply comment.
*
* @param replyCommentDTO the reply comment dto
* @return the integer
*/
|
Creates the reply comment
|
createReplyComment
|
{
"repo_name": "aholake/hiringviet",
"path": "src/main/java/vn/com/hiringviet/service/ReplyCommentService.java",
"license": "apache-2.0",
"size": 969
}
|
[
"vn.com.hiringviet.dto.ReplyCommentDTO"
] |
import vn.com.hiringviet.dto.ReplyCommentDTO;
|
import vn.com.hiringviet.dto.*;
|
[
"vn.com.hiringviet"
] |
vn.com.hiringviet;
| 2,236,051
|
for (final StravaSubscriptionObjectType type : StravaSubscriptionObjectType.values()) {
assertNotNull(type.getDescription());
}
}
|
for (final StravaSubscriptionObjectType type : StravaSubscriptionObjectType.values()) { assertNotNull(type.getDescription()); } }
|
/**
* Test returning the description
*/
|
Test returning the description
|
testGetDescription
|
{
"repo_name": "danshannon/javastrava-test",
"path": "src/main/java/test/model/reference/StravaSubscriptionObjectTypeTest.java",
"license": "apache-2.0",
"size": 887
}
|
[
"org.junit.Assert"
] |
import org.junit.Assert;
|
import org.junit.*;
|
[
"org.junit"
] |
org.junit;
| 2,494,881
|
public void testLocationComparator() {
// Change the GeoLocation of the comments
GeoLocation specificLoc = new GeoLocation(0.0, 0.0);
GeoLocation geoLoc1 = new GeoLocation(1.4, 0.0);
GeoLocation geoLoc2 = new GeoLocation(8.0, 20.0);
GeoLocation geoLoc3 = new GeoLocation(-2.0, -4.0);
// Attach locations to comments
comment1.setLocation(geoLoc1);
comment2.setLocation(geoLoc2);
comment3.setLocation(geoLoc3);
// Create a list of (distance from current location, comment) pairs.
ArrayList<Pair<Double, Comment>> pairs = new ArrayList<Pair<Double, Comment>>();
for(Comment comment: comments) {
Double distance = specificLoc.distanceFrom(comment.getLocation());
pairs.add(Pair.create(distance, comment));
}
Collections.sort(pairs, new LocationComparator());
// Check the comments were sorted by a specific location.
assertTrue("Comments should be sorted by a specific geolocation",
pairs.get(0).first < pairs.get(1).first);
assertTrue(pairs.get(1).first < pairs.get(2).first);
}
|
void function() { GeoLocation specificLoc = new GeoLocation(0.0, 0.0); GeoLocation geoLoc1 = new GeoLocation(1.4, 0.0); GeoLocation geoLoc2 = new GeoLocation(8.0, 20.0); GeoLocation geoLoc3 = new GeoLocation(-2.0, -4.0); comment1.setLocation(geoLoc1); comment2.setLocation(geoLoc2); comment3.setLocation(geoLoc3); ArrayList<Pair<Double, Comment>> pairs = new ArrayList<Pair<Double, Comment>>(); for(Comment comment: comments) { Double distance = specificLoc.distanceFrom(comment.getLocation()); pairs.add(Pair.create(distance, comment)); } Collections.sort(pairs, new LocationComparator()); assertTrue(STR, pairs.get(0).first < pairs.get(1).first); assertTrue(pairs.get(1).first < pairs.get(2).first); }
|
/**
* Checks that comments can be sorted based on some fixed GeoLocation using
* the LocationComparator.
*/
|
Checks that comments can be sorted based on some fixed GeoLocation using the LocationComparator
|
testLocationComparator
|
{
"repo_name": "CMPUT301W14T06/LARD",
"path": "LARDTestTest/src/ca/ualberta/lard/test/ComparatorTests.java",
"license": "mit",
"size": 4501
}
|
[
"android.util.Pair",
"ca.ualberta.lard.comparator.LocationComparator",
"ca.ualberta.lard.model.Comment",
"ca.ualberta.lard.model.GeoLocation",
"java.util.ArrayList",
"java.util.Collections"
] |
import android.util.Pair; import ca.ualberta.lard.comparator.LocationComparator; import ca.ualberta.lard.model.Comment; import ca.ualberta.lard.model.GeoLocation; import java.util.ArrayList; import java.util.Collections;
|
import android.util.*; import ca.ualberta.lard.comparator.*; import ca.ualberta.lard.model.*; import java.util.*;
|
[
"android.util",
"ca.ualberta.lard",
"java.util"
] |
android.util; ca.ualberta.lard; java.util;
| 2,554,880
|
public Integer removeAction(User loggedInUser,
String chainLabel,
Integer actionId) {
ActionChain chain = this.acUtil.getActionChainByLabel(loggedInUser, chainLabel);
for (ActionChainEntry entry : chain.getEntries()) {
if (entry.getAction().getId().equals(Long.valueOf(actionId))) {
ActionChainFactory.removeActionChainEntry(chain, entry);
return BaseHandler.VALID;
}
}
throw new NoSuchActionException("ID: " + actionId);
}
|
Integer function(User loggedInUser, String chainLabel, Integer actionId) { ActionChain chain = this.acUtil.getActionChainByLabel(loggedInUser, chainLabel); for (ActionChainEntry entry : chain.getEntries()) { if (entry.getAction().getId().equals(Long.valueOf(actionId))) { ActionChainFactory.removeActionChainEntry(chain, entry); return BaseHandler.VALID; } } throw new NoSuchActionException(STR + actionId); }
|
/**
* Remove an action from the Action Chain.
*
* @param loggedInUser The current user
* @param chainLabel The label of the Action Chain.
* @param actionId Action ID.
* @return 1 if successful, exception otherwise
*
* @xmlrpc.doc Remove an action from an Action Chain.
* @xmlrpc.param #param_desc("string", "sessionKey", "Session token, issued at login")
* @xmlrpc.param #param_desc("string", "chainLabel", "Label of the chain")
* @xmlrpc.param #param_desc("int", "actionId", "Action ID")
* @xmlrpc.returntype #return_int_success()
*/
|
Remove an action from the Action Chain
|
removeAction
|
{
"repo_name": "xkollar/spacewalk",
"path": "java/code/src/com/redhat/rhn/frontend/xmlrpc/chain/ActionChainHandler.java",
"license": "gpl-2.0",
"size": 21753
}
|
[
"com.redhat.rhn.domain.action.ActionChain",
"com.redhat.rhn.domain.action.ActionChainEntry",
"com.redhat.rhn.domain.action.ActionChainFactory",
"com.redhat.rhn.domain.user.User",
"com.redhat.rhn.frontend.xmlrpc.BaseHandler",
"com.redhat.rhn.frontend.xmlrpc.NoSuchActionException"
] |
import com.redhat.rhn.domain.action.ActionChain; import com.redhat.rhn.domain.action.ActionChainEntry; import com.redhat.rhn.domain.action.ActionChainFactory; import com.redhat.rhn.domain.user.User; import com.redhat.rhn.frontend.xmlrpc.BaseHandler; import com.redhat.rhn.frontend.xmlrpc.NoSuchActionException;
|
import com.redhat.rhn.domain.action.*; import com.redhat.rhn.domain.user.*; import com.redhat.rhn.frontend.xmlrpc.*;
|
[
"com.redhat.rhn"
] |
com.redhat.rhn;
| 2,689,121
|
@Override
public String toString() {
if (isEmpty()) {
return "{}"; //$NON-NLS-1$
}
StringBuilder buffer = new StringBuilder(size() * 28);
buffer.append('{');
Iterator<Map.Entry<K, V>> it = entrySet().iterator();
while (it.hasNext()) {
Map.Entry<K, V> entry = it.next();
Object key = entry.getKey();
if (key != this) {
buffer.append(key);
} else {
buffer.append("(this Map)"); //$NON-NLS-1$
}
buffer.append('=');
Object value = entry.getValue();
if (value != this) {
buffer.append(value);
} else {
buffer.append("(this Map)"); //$NON-NLS-1$
}
if (it.hasNext()) {
buffer.append(", "); //$NON-NLS-1$
}
}
buffer.append('}');
return buffer.toString();
}
/**
* Returns a collection of the values contained in this map. The collection
* is backed by this map so changes to one are reflected by the other. The
* collection supports remove, removeAll, retainAll and clear operations,
* and it does not support add or addAll operations.
* <p>
* This method returns a collection which is the subclass of
* AbstractCollection. The iterator method of this subclass returns a
* "wrapper object" over the iterator of map's entrySet(). The {@code size}
|
String function() { if (isEmpty()) { return "{}"; } StringBuilder buffer = new StringBuilder(size() * 28); buffer.append('{'); Iterator<Map.Entry<K, V>> it = entrySet().iterator(); while (it.hasNext()) { Map.Entry<K, V> entry = it.next(); Object key = entry.getKey(); if (key != this) { buffer.append(key); } else { buffer.append(STR); } buffer.append('='); Object value = entry.getValue(); if (value != this) { buffer.append(value); } else { buffer.append(STR); } if (it.hasNext()) { buffer.append(STR); } } buffer.append('}'); return buffer.toString(); } /** * Returns a collection of the values contained in this map. The collection * is backed by this map so changes to one are reflected by the other. The * collection supports remove, removeAll, retainAll and clear operations, * and it does not support add or addAll operations. * <p> * This method returns a collection which is the subclass of * AbstractCollection. The iterator method of this subclass returns a * STR over the iterator of map's entrySet(). The {@code size}
|
/**
* Returns the string representation of this map.
*
* @return the string representation of this map.
*/
|
Returns the string representation of this map
|
toString
|
{
"repo_name": "Programming-Systems-Lab/phosphor",
"path": "Phosphor/src/main/java/edu/columbia/cs/psl/phosphor/struct/harmony/util/AbstractMap.java",
"license": "mit",
"size": 15170
}
|
[
"java.util.Iterator"
] |
import java.util.Iterator;
|
import java.util.*;
|
[
"java.util"
] |
java.util;
| 1,164,382
|
boolean supports( final Path path );
|
boolean supports( final Path path );
|
/**
* Helpers signal whether it supports the given Path
* @param path
* The Path that was deleted.
* @return
*/
|
Helpers signal whether it supports the given Path
|
supports
|
{
"repo_name": "kiereleaseuser/uberfire",
"path": "uberfire-extensions/uberfire-commons-editor/uberfire-commons-editor-backend/src/main/java/org/uberfire/ext/editor/commons/backend/service/helper/DeleteHelper.java",
"license": "apache-2.0",
"size": 1273
}
|
[
"org.uberfire.backend.vfs.Path"
] |
import org.uberfire.backend.vfs.Path;
|
import org.uberfire.backend.vfs.*;
|
[
"org.uberfire.backend"
] |
org.uberfire.backend;
| 590,290
|
public double getSpecialDoubleProperty(String name,
Hashtable valueSet,
String defaultKey)
{
return this.getDoubleAttribute(name, valueSet, defaultKey, true);
}
|
double function(String name, Hashtable valueSet, String defaultKey) { return this.getDoubleAttribute(name, valueSet, defaultKey, true); }
|
/**
* Returns an attribute by looking up a key in a hashtable.
*
* @deprecated Use {@link #getDoubleAttribute(java.lang.String,
* java.util.Hashtable, java.lang.String, boolean)
* getDoubleAttribute} instead.
*/
|
Returns an attribute by looking up a key in a hashtable
|
getSpecialDoubleProperty
|
{
"repo_name": "lsilvestre/Jogre",
"path": "api/src/nanoxml/XMLElement.java",
"license": "gpl-2.0",
"size": 99514
}
|
[
"java.util.Hashtable"
] |
import java.util.Hashtable;
|
import java.util.*;
|
[
"java.util"
] |
java.util;
| 1,896,477
|
private static String getClassName(String fullClassName) {
if(!StringUtil.isSet(fullClassName)) {
return null;
}
fullClassName += " ";
char[] classNameArray = fullClassName.toCharArray();
StringBuilder buffer = new StringBuilder();
List<String> tokens = new ArrayList<String>();
String className;
int dotIndex;
for(int i = 0; i < fullClassName.length(); ++i) {
if("<>, ".indexOf(classNameArray[i]) > -1) {
if(buffer.length() > 0) {
className = buffer.toString();
dotIndex = className.lastIndexOf('.');
if(dotIndex > -1) {
className = className.substring(dotIndex+1);
}
tokens.add(className.trim());
}
buffer = new StringBuilder();
tokens.add(classNameArray[i] + "");
}
else {
buffer.append(classNameArray[i]);
}
}
StringBuilder result = new StringBuilder();
for (String string : tokens) {
result.append(string);
}
return result.toString().trim();
}
|
static String function(String fullClassName) { if(!StringUtil.isSet(fullClassName)) { return null; } fullClassName += " "; char[] classNameArray = fullClassName.toCharArray(); StringBuilder buffer = new StringBuilder(); List<String> tokens = new ArrayList<String>(); String className; int dotIndex; for(int i = 0; i < fullClassName.length(); ++i) { if(STR.indexOf(classNameArray[i]) > -1) { if(buffer.length() > 0) { className = buffer.toString(); dotIndex = className.lastIndexOf('.'); if(dotIndex > -1) { className = className.substring(dotIndex+1); } tokens.add(className.trim()); } buffer = new StringBuilder(); tokens.add(classNameArray[i] + ""); } else { buffer.append(classNameArray[i]); } } StringBuilder result = new StringBuilder(); for (String string : tokens) { result.append(string); } return result.toString().trim(); }
|
/**
* Converts a full class name into just the class name.
*
* @param fullClassName
* an objects fully qualified class name.
* @return The class name of the object with the package portion removed.
*/
|
Converts a full class name into just the class name
|
getClassName
|
{
"repo_name": "justinrknowles/toshookan",
"path": "src/main/java/toshookan/domainmodel/VelocityDomainModelCodeGenerator.java",
"license": "apache-2.0",
"size": 16799
}
|
[
"java.util.ArrayList",
"java.util.List"
] |
import java.util.ArrayList; import java.util.List;
|
import java.util.*;
|
[
"java.util"
] |
java.util;
| 1,000,775
|
@Override
public Spans getSpans(final LeafReaderContext context, final Bits acceptDocs, final Map<Term,TermContext> termContexts) throws IOException {
ArrayList<Spans> containerContained = prepareConjunction(context, acceptDocs, termContexts);
if (containerContained == null) {
return null;
}
Spans big = containerContained.get(0);
Spans little = containerContained.get(1);
return new ContainSpans(big, little, little) {
|
Spans function(final LeafReaderContext context, final Bits acceptDocs, final Map<Term,TermContext> termContexts) throws IOException { ArrayList<Spans> containerContained = prepareConjunction(context, acceptDocs, termContexts); if (containerContained == null) { return null; } Spans big = containerContained.get(0); Spans little = containerContained.get(1); return new ContainSpans(big, little, little) {
|
/**
* Return spans from <code>little</code> that are contained in a spans from <code>big</code>.
* The payload is from the spans of <code>little</code>.
*/
|
Return spans from <code>little</code> that are contained in a spans from <code>big</code>. The payload is from the spans of <code>little</code>
|
getSpans
|
{
"repo_name": "q474818917/solr-5.2.0",
"path": "lucene/core/src/java/org/apache/lucene/search/spans/SpanWithinQuery.java",
"license": "apache-2.0",
"size": 3931
}
|
[
"java.io.IOException",
"java.util.ArrayList",
"java.util.Map",
"org.apache.lucene.index.LeafReaderContext",
"org.apache.lucene.index.Term",
"org.apache.lucene.index.TermContext",
"org.apache.lucene.util.Bits"
] |
import java.io.IOException; import java.util.ArrayList; import java.util.Map; import org.apache.lucene.index.LeafReaderContext; import org.apache.lucene.index.Term; import org.apache.lucene.index.TermContext; import org.apache.lucene.util.Bits;
|
import java.io.*; import java.util.*; import org.apache.lucene.index.*; import org.apache.lucene.util.*;
|
[
"java.io",
"java.util",
"org.apache.lucene"
] |
java.io; java.util; org.apache.lucene;
| 2,819,000
|
public Options validateIndices(Boolean validateIndices) {
this.validateIndices = validateIndices;
return this;
}
}
@OpInputsMetadata(
outputsClass = DenseToDenseSetOperation.class
)
public static class Inputs<T extends TType> extends RawOpInputs<DenseToDenseSetOperation<T>> {
public final Operand<T> set1;
public final Operand<T> set2;
public final String setOperation;
public final boolean validateIndices;
public final DataType T;
public Inputs(GraphOperation op) {
super(new DenseToDenseSetOperation<>(op), op, Arrays.asList("set_operation", "validate_indices", "T"));
int inputIndex = 0;
set1 = (Operand<T>) op.input(inputIndex++);
set2 = (Operand<T>) op.input(inputIndex++);
setOperation = op.attributes().getAttrString("set_operation");
validateIndices = op.attributes().getAttrBool("validate_indices");
T = op.attributes().getAttrType("T");
}
}
|
Options function(Boolean validateIndices) { this.validateIndices = validateIndices; return this; } } @OpInputsMetadata( outputsClass = DenseToDenseSetOperation.class ) static class Inputs<T extends TType> extends RawOpInputs<DenseToDenseSetOperation<T>> { public final Operand<T> set1; public final Operand<T> set2; public final String setOperation; public final boolean function; final DataType T; public Inputs(GraphOperation op) { super(new DenseToDenseSetOperation<>(op), op, Arrays.asList(STR, STR, "T")); int inputIndex = 0; set1 = (Operand<T>) op.input(inputIndex++); set2 = (Operand<T>) op.input(inputIndex++); setOperation = op.attributes().getAttrString(STR); function = op.attributes().getAttrBool(STR); T = op.attributes().getAttrType("T"); } }
|
/**
* Sets the validateIndices option.
*
* @param validateIndices the validateIndices option
* @return this Options instance.
*/
|
Sets the validateIndices option
|
validateIndices
|
{
"repo_name": "tensorflow/java",
"path": "tensorflow-core/tensorflow-core-api/src/gen/java/org/tensorflow/op/sparse/DenseToDenseSetOperation.java",
"license": "apache-2.0",
"size": 7203
}
|
[
"java.util.Arrays",
"org.tensorflow.GraphOperation",
"org.tensorflow.Operand",
"org.tensorflow.op.RawOpInputs",
"org.tensorflow.op.annotation.OpInputsMetadata",
"org.tensorflow.proto.framework.DataType",
"org.tensorflow.types.family.TType"
] |
import java.util.Arrays; import org.tensorflow.GraphOperation; import org.tensorflow.Operand; import org.tensorflow.op.RawOpInputs; import org.tensorflow.op.annotation.OpInputsMetadata; import org.tensorflow.proto.framework.DataType; import org.tensorflow.types.family.TType;
|
import java.util.*; import org.tensorflow.*; import org.tensorflow.op.*; import org.tensorflow.op.annotation.*; import org.tensorflow.proto.framework.*; import org.tensorflow.types.family.*;
|
[
"java.util",
"org.tensorflow",
"org.tensorflow.op",
"org.tensorflow.proto",
"org.tensorflow.types"
] |
java.util; org.tensorflow; org.tensorflow.op; org.tensorflow.proto; org.tensorflow.types;
| 1,187,165
|
Set<LogListener> getLogListeners();
|
Set<LogListener> getLogListeners();
|
/**
* Gets a list of {@link LogListener}.
*/
|
Gets a list of <code>LogListener</code>
|
getLogListeners
|
{
"repo_name": "tlehoux/camel",
"path": "camel-core/src/main/java/org/apache/camel/CamelContext.java",
"license": "apache-2.0",
"size": 77596
}
|
[
"java.util.Set",
"org.apache.camel.spi.LogListener"
] |
import java.util.Set; import org.apache.camel.spi.LogListener;
|
import java.util.*; import org.apache.camel.spi.*;
|
[
"java.util",
"org.apache.camel"
] |
java.util; org.apache.camel;
| 1,646,593
|
@Test ( expected = PicturesComparator.PicturesComparatorException.class )
public void comparePicturesWithNullPickedPicture() throws PicturesComparator.PicturesComparatorException {
l(this, "@Test comparePicturesWithNullPickedPicture");
PicturesComparator pc = new PixelByPixelPicturesComparator();
pc.comparePictures(createEmptyBitmap(), null);
}
|
@Test ( expected = PicturesComparator.PicturesComparatorException.class ) void function() throws PicturesComparator.PicturesComparatorException { l(this, STR); PicturesComparator pc = new PixelByPixelPicturesComparator(); pc.comparePictures(createEmptyBitmap(), null); }
|
/**
* Tests the comparePictures() with null picked picture
*
* <i>Comparisons with null picked picture must thrown an exception</i>
*/
|
Tests the comparePictures() with null picked picture Comparisons with null picked picture must thrown an exception
|
comparePicturesWithNullPickedPicture
|
{
"repo_name": "pylapp/SmoothClicker",
"path": "app/app/src/androidTest/java/pylapp/smoothclicker/android/tools/screen/ItPixelByPixelPicturesComparator.java",
"license": "mit",
"size": 12817
}
|
[
"org.junit.Test"
] |
import org.junit.Test;
|
import org.junit.*;
|
[
"org.junit"
] |
org.junit;
| 1,104,951
|
personalDictionary = new PersonalDictionary(personalFile, dictionaryEngine.get_dic_encoding()) {
protected void dictionaryChanged() {
super.dictionaryChanged();
fireDictionaryConfigurationChanged();
}
};
// Fill-in the dictionary engine
List words = personalDictionary.getWords();
for (Iterator iter = words.iterator(); iter.hasNext();) {
String customWord = (String) iter.next();
dictionaryEngine.addCustomWord(customWord);
}
}
|
personalDictionary = new PersonalDictionary(personalFile, dictionaryEngine.get_dic_encoding()) { void function() { super.dictionaryChanged(); fireDictionaryConfigurationChanged(); } }; List words = personalDictionary.getWords(); for (Iterator iter = words.iterator(); iter.hasNext();) { String customWord = (String) iter.next(); dictionaryEngine.addCustomWord(customWord); } }
|
/**
* ensure that any direct change to the PersonalDictionary fires an event to the main dictionary
*/
|
ensure that any direct change to the PersonalDictionary fires an event to the main dictionary
|
dictionaryChanged
|
{
"repo_name": "dbaeli/spell6r",
"path": "src/main/java/org/dts/spell/dictionary/openoffice/OpenOfficeSpellDictionary.java",
"license": "lgpl-2.1",
"size": 5802
}
|
[
"java.util.Iterator",
"java.util.List"
] |
import java.util.Iterator; import java.util.List;
|
import java.util.*;
|
[
"java.util"
] |
java.util;
| 2,226,243
|
public boolean isEventBelongingToNode(GUIMouseEvent event) {
return isEventBelongingToNode(event, null);
}
|
boolean function(GUIMouseEvent event) { return isEventBelongingToNode(event, null); }
|
/**
* Is event belonging to node
* @param event
* @return boolean
*/
|
Is event belonging to node
|
isEventBelongingToNode
|
{
"repo_name": "andreasdr/tdme",
"path": "src/net/drewke/tdme/gui/nodes/GUINode.java",
"license": "mit",
"size": 35207
}
|
[
"net.drewke.tdme.gui.events.GUIMouseEvent"
] |
import net.drewke.tdme.gui.events.GUIMouseEvent;
|
import net.drewke.tdme.gui.events.*;
|
[
"net.drewke.tdme"
] |
net.drewke.tdme;
| 378,920
|
public static JavadocTags getJavadocTags(TextBlock cmt,
JavadocTagType tagType) {
final String[] text = cmt.getText();
final List<JavadocTag> tags = Lists.newArrayList();
final List<InvalidJavadocTag> invalidTags = Lists.newArrayList();
Pattern blockTagPattern = Pattern.compile("/\\*{2,}\\s*@(\\p{Alpha}+)\\s");
for (int i = 0; i < text.length; i++) {
final String textValue = text[i];
final Matcher blockTagMatcher = blockTagPattern.matcher(textValue);
if ((tagType == JavadocTagType.ALL || tagType == JavadocTagType.BLOCK)
&& blockTagMatcher.find()) {
final String tagName = blockTagMatcher.group(1);
String content = textValue.substring(blockTagMatcher.end(1));
if (content.endsWith("*/")) {
content = content.substring(0, content.length() - 2);
}
final int line = cmt.getStartLineNo() + i;
int col = blockTagMatcher.start(1) - 1;
if (i == 0) {
col += cmt.getStartColNo();
}
if (JavadocTagInfo.isValidName(tagName)) {
tags.add(
new JavadocTag(line, col, tagName, content.trim()));
}
else {
invalidTags.add(new InvalidJavadocTag(line, col, tagName));
}
}
// No block tag, so look for inline validTags
else if (tagType == JavadocTagType.ALL || tagType == JavadocTagType.INLINE) {
lookForInlineTags(cmt, i, tags, invalidTags);
}
blockTagPattern = Pattern.compile("^\\s*\\**\\s*@(\\p{Alpha}+)\\s");
}
return new JavadocTags(tags, invalidTags);
}
|
static JavadocTags function(TextBlock cmt, JavadocTagType tagType) { final String[] text = cmt.getText(); final List<JavadocTag> tags = Lists.newArrayList(); final List<InvalidJavadocTag> invalidTags = Lists.newArrayList(); Pattern blockTagPattern = Pattern.compile(STR); for (int i = 0; i < text.length; i++) { final String textValue = text[i]; final Matcher blockTagMatcher = blockTagPattern.matcher(textValue); if ((tagType == JavadocTagType.ALL tagType == JavadocTagType.BLOCK) && blockTagMatcher.find()) { final String tagName = blockTagMatcher.group(1); String content = textValue.substring(blockTagMatcher.end(1)); if (content.endsWith("*/")) { content = content.substring(0, content.length() - 2); } final int line = cmt.getStartLineNo() + i; int col = blockTagMatcher.start(1) - 1; if (i == 0) { col += cmt.getStartColNo(); } if (JavadocTagInfo.isValidName(tagName)) { tags.add( new JavadocTag(line, col, tagName, content.trim())); } else { invalidTags.add(new InvalidJavadocTag(line, col, tagName)); } } else if (tagType == JavadocTagType.ALL tagType == JavadocTagType.INLINE) { lookForInlineTags(cmt, i, tags, invalidTags); } blockTagPattern = Pattern.compile(STR); } return new JavadocTags(tags, invalidTags); }
|
/**
* Gets validTags from a given piece of Javadoc.
* @param cmt
* the Javadoc comment to process.
* @param tagType
* the type of validTags we're interested in
* @return all standalone validTags from the given javadoc.
*/
|
Gets validTags from a given piece of Javadoc
|
getJavadocTags
|
{
"repo_name": "gallandarakhneorg/checkstyle",
"path": "src/main/java/com/puppycrawl/tools/checkstyle/utils/JavadocUtils.java",
"license": "lgpl-2.1",
"size": 15258
}
|
[
"com.google.common.collect.Lists",
"com.puppycrawl.tools.checkstyle.api.TextBlock",
"com.puppycrawl.tools.checkstyle.checks.javadoc.InvalidJavadocTag",
"com.puppycrawl.tools.checkstyle.checks.javadoc.JavadocTag",
"com.puppycrawl.tools.checkstyle.checks.javadoc.JavadocTagInfo",
"com.puppycrawl.tools.checkstyle.checks.javadoc.JavadocTags",
"java.util.List",
"java.util.regex.Matcher",
"java.util.regex.Pattern"
] |
import com.google.common.collect.Lists; import com.puppycrawl.tools.checkstyle.api.TextBlock; import com.puppycrawl.tools.checkstyle.checks.javadoc.InvalidJavadocTag; import com.puppycrawl.tools.checkstyle.checks.javadoc.JavadocTag; import com.puppycrawl.tools.checkstyle.checks.javadoc.JavadocTagInfo; import com.puppycrawl.tools.checkstyle.checks.javadoc.JavadocTags; import java.util.List; import java.util.regex.Matcher; import java.util.regex.Pattern;
|
import com.google.common.collect.*; import com.puppycrawl.tools.checkstyle.api.*; import com.puppycrawl.tools.checkstyle.checks.javadoc.*; import java.util.*; import java.util.regex.*;
|
[
"com.google.common",
"com.puppycrawl.tools",
"java.util"
] |
com.google.common; com.puppycrawl.tools; java.util;
| 2,395,545
|
public static ApprovalDialog getInformationDialog(Dialog owner, ModalityType modal) {
ApprovalDialog result;
result = new ApprovalDialog(owner, modal);
result.setApproveVisible(true);
result.setDiscardVisible(false);
result.setCancelVisible(false);
return result;
}
|
static ApprovalDialog function(Dialog owner, ModalityType modal) { ApprovalDialog result; result = new ApprovalDialog(owner, modal); result.setApproveVisible(true); result.setDiscardVisible(false); result.setCancelVisible(false); return result; }
|
/**
* Returns a basic info dialog (ok).
*
* @param owner the owner of the dialog
* @param modal the modality of the dialog
*/
|
Returns a basic info dialog (ok)
|
getInformationDialog
|
{
"repo_name": "Waikato/fcms-widgets",
"path": "src/main/java/nz/ac/waikato/cms/gui/core/ApprovalDialog.java",
"license": "gpl-3.0",
"size": 14246
}
|
[
"java.awt.Dialog"
] |
import java.awt.Dialog;
|
import java.awt.*;
|
[
"java.awt"
] |
java.awt;
| 96,819
|
public String getPicturePosition() throws SonyProjectorException {
if (!model.isPicturePositionAvailable()) {
throw new SonyProjectorException("Unavailable item " + SonyProjectorItem.PICTURE_POSITION.getName()
+ " for projector model " + model.getName());
}
return model.getPicturePositionNameFromDataCode(getSetting(SonyProjectorItem.PICTURE_POSITION));
}
|
String function() throws SonyProjectorException { if (!model.isPicturePositionAvailable()) { throw new SonyProjectorException(STR + SonyProjectorItem.PICTURE_POSITION.getName() + STR + model.getName()); } return model.getPicturePositionNameFromDataCode(getSetting(SonyProjectorItem.PICTURE_POSITION)); }
|
/**
* Request the projector to get the current mode for the picture position setting
*
* @return the current mode for the picture position setting
*
* @throws SonyProjectorException - In case this setting is not available for the projector or any other problem
*/
|
Request the projector to get the current mode for the picture position setting
|
getPicturePosition
|
{
"repo_name": "openhab/openhab2",
"path": "bundles/org.openhab.binding.sonyprojector/src/main/java/org/openhab/binding/sonyprojector/internal/communication/SonyProjectorConnector.java",
"license": "epl-1.0",
"size": 43215
}
|
[
"org.openhab.binding.sonyprojector.internal.SonyProjectorException"
] |
import org.openhab.binding.sonyprojector.internal.SonyProjectorException;
|
import org.openhab.binding.sonyprojector.internal.*;
|
[
"org.openhab.binding"
] |
org.openhab.binding;
| 93,698
|
public Locale getLocale() {
return m_locale;
}
|
Locale function() { return m_locale; }
|
/**
* Returns the locale.<p>
*
* @return the locale
*/
|
Returns the locale
|
getLocale
|
{
"repo_name": "alkacon/opencms-core",
"path": "src/org/opencms/xml/content/CmsMappingResolutionContext.java",
"license": "lgpl-2.1",
"size": 11806
}
|
[
"java.util.Locale"
] |
import java.util.Locale;
|
import java.util.*;
|
[
"java.util"
] |
java.util;
| 2,909,839
|
NetSuiteConnectionProperties getConnectionProperties();
/**
* Return identifier of referenced connection component.
*
* @return referenced connection component's ID or {@code null}
|
NetSuiteConnectionProperties getConnectionProperties(); /** * Return identifier of referenced connection component. * * @return referenced connection component's ID or {@code null}
|
/**
* Return connection properties used by this properties object.
*
* @return connection properties
*/
|
Return connection properties used by this properties object
|
getConnectionProperties
|
{
"repo_name": "Talend/components",
"path": "components/components-netsuite/components-netsuite-definition/src/main/java/org/talend/components/netsuite/NetSuiteProvideConnectionProperties.java",
"license": "apache-2.0",
"size": 1181
}
|
[
"org.talend.components.netsuite.connection.NetSuiteConnectionProperties"
] |
import org.talend.components.netsuite.connection.NetSuiteConnectionProperties;
|
import org.talend.components.netsuite.connection.*;
|
[
"org.talend.components"
] |
org.talend.components;
| 1,126,876
|
public GitRepository createBareRepo(GeneralOptions generalOptions, Path path)
throws RepoException {
GitRepository repo =
GitRepository.newBareRepo(
path, getGitEnvironment(generalOptions.getEnvironment()),
generalOptions.isVerbose(), generalOptions.fetchTimeout, gitNoVerify);
return initRepo(repo);
}
|
GitRepository function(GeneralOptions generalOptions, Path path) throws RepoException { GitRepository repo = GitRepository.newBareRepo( path, getGitEnvironment(generalOptions.getEnvironment()), generalOptions.isVerbose(), generalOptions.fetchTimeout, gitNoVerify); return initRepo(repo); }
|
/**
* Create a new initialized repository in the location.
*
* <p>Can be overwritten to create custom GitRepository objects.
*/
|
Create a new initialized repository in the location. Can be overwritten to create custom GitRepository objects
|
createBareRepo
|
{
"repo_name": "google/copybara",
"path": "java/com/google/copybara/git/GitOptions.java",
"license": "apache-2.0",
"size": 5466
}
|
[
"com.google.copybara.GeneralOptions",
"com.google.copybara.exception.RepoException",
"java.nio.file.Path"
] |
import com.google.copybara.GeneralOptions; import com.google.copybara.exception.RepoException; import java.nio.file.Path;
|
import com.google.copybara.*; import com.google.copybara.exception.*; import java.nio.file.*;
|
[
"com.google.copybara",
"java.nio"
] |
com.google.copybara; java.nio;
| 143,831
|
private void logTimeoutObjectsFrequency() {
StringBuilder sb = new StringBuilder("Timeout objects frequency [");
for (Ignite ignite : G.allGrids()) {
IgniteEx igniteEx = (IgniteEx)ignite;
Map<String, Integer> objFreqMap = new HashMap<>();
Set<GridTimeoutObject> objs = getTimeoutObjects(igniteEx);
for (GridTimeoutObject obj : objs) {
String clsName = obj.getClass().getSimpleName();
Integer cnt = objFreqMap.get(clsName);
if (cnt == null)
objFreqMap.put(clsName, 1);
else
objFreqMap.put(clsName, cnt + 1);
}
sb.append("[")
.append(igniteEx.name()).append(": size=")
.append(objs.size()).append(", ");
for (Map.Entry<String, Integer> entry : objFreqMap.entrySet()) {
sb.append(entry.getKey()).append("=")
.append(entry.getValue())
.append(", ");
}
sb.delete(sb.length() - 2, sb.length())
.append("]; ");
}
sb.delete(sb.length() - 2, sb.length())
.append("]");
info(sb.toString()
.replaceAll("distributed.IgniteTxRemoveTimeoutObjectsTest", "Grid"));
}
|
void function() { StringBuilder sb = new StringBuilder(STR); for (Ignite ignite : G.allGrids()) { IgniteEx igniteEx = (IgniteEx)ignite; Map<String, Integer> objFreqMap = new HashMap<>(); Set<GridTimeoutObject> objs = getTimeoutObjects(igniteEx); for (GridTimeoutObject obj : objs) { String clsName = obj.getClass().getSimpleName(); Integer cnt = objFreqMap.get(clsName); if (cnt == null) objFreqMap.put(clsName, 1); else objFreqMap.put(clsName, cnt + 1); } sb.append("[") .append(igniteEx.name()).append(STR) .append(objs.size()).append(STR); for (Map.Entry<String, Integer> entry : objFreqMap.entrySet()) { sb.append(entry.getKey()).append("=") .append(entry.getValue()) .append(STR); } sb.delete(sb.length() - 2, sb.length()) .append(STR); } sb.delete(sb.length() - 2, sb.length()) .append("]"); info(sb.toString() .replaceAll("distributed.IgniteTxRemoveTimeoutObjectsTestSTRGrid")); }
|
/**
* Print the number of each timeout object type on each grid to the log.
*/
|
Print the number of each timeout object type on each grid to the log
|
logTimeoutObjectsFrequency
|
{
"repo_name": "vladisav/ignite",
"path": "modules/core/src/test/java/org/apache/ignite/internal/processors/cache/distributed/IgniteTxRemoveTimeoutObjectsTest.java",
"license": "apache-2.0",
"size": 7168
}
|
[
"java.util.HashMap",
"java.util.Map",
"java.util.Set",
"org.apache.ignite.Ignite",
"org.apache.ignite.internal.IgniteEx",
"org.apache.ignite.internal.processors.timeout.GridTimeoutObject",
"org.apache.ignite.internal.util.typedef.G"
] |
import java.util.HashMap; import java.util.Map; import java.util.Set; import org.apache.ignite.Ignite; import org.apache.ignite.internal.IgniteEx; import org.apache.ignite.internal.processors.timeout.GridTimeoutObject; import org.apache.ignite.internal.util.typedef.G;
|
import java.util.*; import org.apache.ignite.*; import org.apache.ignite.internal.*; import org.apache.ignite.internal.processors.timeout.*; import org.apache.ignite.internal.util.typedef.*;
|
[
"java.util",
"org.apache.ignite"
] |
java.util; org.apache.ignite;
| 2,304,515
|
private String getLocalizedKey(String propertyKey,
boolean useDefaultLanguage) {
String language;
if (useDefaultLanguage) {
language = defaultLocale.getLanguage();
} else {
Locale locale = Locale.getDefault();
language = locale.getLanguage();
}
return propertyKey + "_" + language;
}
|
String function(String propertyKey, boolean useDefaultLanguage) { String language; if (useDefaultLanguage) { language = defaultLocale.getLanguage(); } else { Locale locale = Locale.getDefault(); language = locale.getLanguage(); } return propertyKey + "_" + language; }
|
/**
* Small helper to retrieve the "localized" property key. E.g. key:
* MAIL_SERVER ==> localized version: MAIL_SERVER_en
*
* @param propertyKey
* the property key to be localized.
* @param useDefaultLanguage
* <code>pass true to use the default locale.</code>
* @return the "localized" version of the property key.
*/
|
Small helper to retrieve the "localized" property key. E.g. key: MAIL_SERVER ==> localized version: MAIL_SERVER_en
|
getLocalizedKey
|
{
"repo_name": "opetrovski/development",
"path": "oscm-devruntime/javasrc/org/oscm/setup/UserNotificationHandler.java",
"license": "apache-2.0",
"size": 18530
}
|
[
"java.util.Locale"
] |
import java.util.Locale;
|
import java.util.*;
|
[
"java.util"
] |
java.util;
| 70,198
|
public int onlineConsistencyRepair() throws IOException, KeeperException,
InterruptedException {
clearState();
LOG.info("Loading regionsinfo from the .META. table");
boolean success = loadMetaEntries();
if (!success) return -1;
// Check if .META. is found only once and in the right place
if (!checkMetaRegion()) {
// Will remove later if we can fix it
errors.reportError("Encountered fatal error. Exiting...");
return -2;
}
// get a list of all tables that have not changed recently.
if (!checkMetaOnly) {
reportTablesInFlux();
}
// get regions according to what is online on each RegionServer
loadDeployedRegions();
// load regiondirs and regioninfos from HDFS
loadHdfsRegionDirs();
loadHdfsRegionInfos();
// Empty cells in .META.?
reportEmptyMetaCells();
// Get disabled tables from ZooKeeper
loadDisabledTables();
// fix the orphan tables
fixOrphanTables();
// Check and fix consistency
checkAndFixConsistency();
// Check integrity (does not fix)
checkIntegrity();
return errors.getErrorList().size();
}
|
int function() throws IOException, KeeperException, InterruptedException { clearState(); LOG.info(STR); boolean success = loadMetaEntries(); if (!success) return -1; if (!checkMetaRegion()) { errors.reportError(STR); return -2; } if (!checkMetaOnly) { reportTablesInFlux(); } loadDeployedRegions(); loadHdfsRegionDirs(); loadHdfsRegionInfos(); reportEmptyMetaCells(); loadDisabledTables(); fixOrphanTables(); checkAndFixConsistency(); checkIntegrity(); return errors.getErrorList().size(); }
|
/**
* This repair method requires the cluster to be online since it contacts
* region servers and the masters. It makes each region's state in HDFS, in
* .META., and deployments consistent.
*
* @return If > 0 , number of errors detected, if < 0 there was an unrecoverable
* error. If 0, we have a clean hbase.
*/
|
This repair method requires the cluster to be online since it contacts region servers and the masters. It makes each region's state in HDFS, in .META., and deployments consistent
|
onlineConsistencyRepair
|
{
"repo_name": "matteobertozzi/hbase",
"path": "hbase-server/src/main/java/org/apache/hadoop/hbase/util/HBaseFsck.java",
"license": "apache-2.0",
"size": 130110
}
|
[
"java.io.IOException",
"org.apache.zookeeper.KeeperException"
] |
import java.io.IOException; import org.apache.zookeeper.KeeperException;
|
import java.io.*; import org.apache.zookeeper.*;
|
[
"java.io",
"org.apache.zookeeper"
] |
java.io; org.apache.zookeeper;
| 738,867
|
@Override public void exitClauseArgs(@NotNull ErlangParser.ClauseArgsContext ctx) { }
|
@Override public void exitClauseArgs(@NotNull ErlangParser.ClauseArgsContext ctx) { }
|
/**
* {@inheritDoc}
*
* <p>The default implementation does nothing.</p>
*/
|
The default implementation does nothing
|
enterClauseArgs
|
{
"repo_name": "IsThisThePayneResidence/intellidots",
"path": "src/main/java/ua/edu/hneu/ast/parsers/ErlangBaseListener.java",
"license": "gpl-3.0",
"size": 35359
}
|
[
"org.antlr.v4.runtime.misc.NotNull"
] |
import org.antlr.v4.runtime.misc.NotNull;
|
import org.antlr.v4.runtime.misc.*;
|
[
"org.antlr.v4"
] |
org.antlr.v4;
| 559,137
|
@Test
public void errorMessageTest9() throws PcepParseException, PcepOutOfBoundMessageException {
byte[] errorMsg = new byte[]{0x20, 0x06, 0x00, 0x14, // common header
0x0D, 0x10, 0x00, 0x08, // PCEP-ERROR Object Header
0x00, 0x00, 0x01, 0x01, 0x0D, 0x10, 0x00, 0x08, // PCEP-ERROR Object Header
0x00, 0x00, 0x01, 0x01};
ChannelBuffer buffer = ChannelBuffers.dynamicBuffer();
buffer.writeBytes(errorMsg);
PcepMessageReader<PcepMessage> reader = PcepFactories.getGenericReader();
PcepMessage message = null;
message = reader.readFrom(buffer);
byte[] testErrorMsg = {0};
ChannelBuffer buf = ChannelBuffers.dynamicBuffer();
assertThat(message, instanceOf(PcepErrorMsg.class));
message.writeTo(buf);
int iReadLen = buf.writerIndex();
testErrorMsg = new byte[iReadLen];
buf.readBytes(testErrorMsg, 0, iReadLen);
assertThat(testErrorMsg, is(errorMsg));
}
|
void function() throws PcepParseException, PcepOutOfBoundMessageException { byte[] errorMsg = new byte[]{0x20, 0x06, 0x00, 0x14, 0x0D, 0x10, 0x00, 0x08, 0x00, 0x00, 0x01, 0x01, 0x0D, 0x10, 0x00, 0x08, 0x00, 0x00, 0x01, 0x01}; ChannelBuffer buffer = ChannelBuffers.dynamicBuffer(); buffer.writeBytes(errorMsg); PcepMessageReader<PcepMessage> reader = PcepFactories.getGenericReader(); PcepMessage message = null; message = reader.readFrom(buffer); byte[] testErrorMsg = {0}; ChannelBuffer buf = ChannelBuffers.dynamicBuffer(); assertThat(message, instanceOf(PcepErrorMsg.class)); message.writeTo(buf); int iReadLen = buf.writerIndex(); testErrorMsg = new byte[iReadLen]; buf.readBytes(testErrorMsg, 0, iReadLen); assertThat(testErrorMsg, is(errorMsg)); }
|
/**
* This test case checks for
* PCEP-ERROR Object, PCEP-ERROR Object
* in PcepErrorMsg message.
*/
|
This test case checks for PCEP-ERROR Object, PCEP-ERROR Object in PcepErrorMsg message
|
errorMessageTest9
|
{
"repo_name": "kuujo/onos",
"path": "protocols/pcep/pcepio/src/test/java/org/onosproject/pcepio/protocol/PcepErrorMsgTest.java",
"license": "apache-2.0",
"size": 29749
}
|
[
"org.hamcrest.MatcherAssert",
"org.hamcrest.Matchers",
"org.hamcrest.core.Is",
"org.jboss.netty.buffer.ChannelBuffer",
"org.jboss.netty.buffer.ChannelBuffers",
"org.onosproject.pcepio.exceptions.PcepOutOfBoundMessageException",
"org.onosproject.pcepio.exceptions.PcepParseException"
] |
import org.hamcrest.MatcherAssert; import org.hamcrest.Matchers; import org.hamcrest.core.Is; import org.jboss.netty.buffer.ChannelBuffer; import org.jboss.netty.buffer.ChannelBuffers; import org.onosproject.pcepio.exceptions.PcepOutOfBoundMessageException; import org.onosproject.pcepio.exceptions.PcepParseException;
|
import org.hamcrest.*; import org.hamcrest.core.*; import org.jboss.netty.buffer.*; import org.onosproject.pcepio.exceptions.*;
|
[
"org.hamcrest",
"org.hamcrest.core",
"org.jboss.netty",
"org.onosproject.pcepio"
] |
org.hamcrest; org.hamcrest.core; org.jboss.netty; org.onosproject.pcepio;
| 531,580
|
final public Collection<Enrolment> getPropaedeuticEnrolments() {
final Collection<Enrolment> result = new ArrayList<Enrolment>();
for (final Enrolment enrolment : getEnrolmentsSet()) {
if (enrolment.isPropaedeutic()) {
result.add(enrolment);
}
}
return result;
}
|
final Collection<Enrolment> function() { final Collection<Enrolment> result = new ArrayList<Enrolment>(); for (final Enrolment enrolment : getEnrolmentsSet()) { if (enrolment.isPropaedeutic()) { result.add(enrolment); } } return result; }
|
/**
* Note that this method must not use the ExtraCurriculumGroup due to the
* pre-Bolonha SCPs
* @return get propaedeutic enrolments
*/
|
Note that this method must not use the ExtraCurriculumGroup due to the pre-Bolonha SCPs
|
getPropaedeuticEnrolments
|
{
"repo_name": "gil-l/fenix",
"path": "src/main/java/org/fenixedu/academic/domain/StudentCurricularPlan.java",
"license": "lgpl-3.0",
"size": 115072
}
|
[
"java.util.ArrayList",
"java.util.Collection"
] |
import java.util.ArrayList; import java.util.Collection;
|
import java.util.*;
|
[
"java.util"
] |
java.util;
| 1,602,798
|
public static Translation[] getTranslations(
final String translations_file) throws Exception
{
final ArrayList<Translation> trans_arr = new ArrayList<Translation>();
// Open the file
final FileInputStream fstream = new FileInputStream(translations_file);
try {
// Convert file into buffered reader which can read line-by-line
final DataInputStream in = new DataInputStream(fstream);
final BufferedReader br = new BufferedReader(new InputStreamReader(in));
String strLine;
int countLines = 0;
// Read File Line By Line
while ((strLine = br.readLine()) != null)
{
// Count lines read from the file
countLines++;
// Remove spaces
strLine = strLine.trim();
// Skip comments
if (strLine.length() <= 0 ||
strLine.startsWith("#"))
continue;
// Expect some_regular_expression_pattern = translation
final int separator = strLine.indexOf("=");
if (separator < 0) {
br.close();
throw new Exception("Missing separator in line "
+ countLines);
}
// Add pattern & replacement to array of translations
final String pattern = strLine.substring(0, separator);
final String replacement = strLine.substring(separator + 1,
strLine.length());
trans_arr.add(new Translation(pattern, replacement));
}
br.close();
}
finally
{
// Close the input stream
fstream.close();
}
// Convert array list into plain array
return trans_arr.toArray(new Translation[trans_arr.size()]);
}
|
static Translation[] function( final String translations_file) throws Exception { final ArrayList<Translation> trans_arr = new ArrayList<Translation>(); final FileInputStream fstream = new FileInputStream(translations_file); try { final DataInputStream in = new DataInputStream(fstream); final BufferedReader br = new BufferedReader(new InputStreamReader(in)); String strLine; int countLines = 0; while ((strLine = br.readLine()) != null) { countLines++; strLine = strLine.trim(); if (strLine.length() <= 0 strLine.startsWith("#")) continue; final int separator = strLine.indexOf("="); if (separator < 0) { br.close(); throw new Exception(STR + countLines); } final String pattern = strLine.substring(0, separator); final String replacement = strLine.substring(separator + 1, strLine.length()); trans_arr.add(new Translation(pattern, replacement)); } br.close(); } finally { fstream.close(); } return trans_arr.toArray(new Translation[trans_arr.size()]); }
|
/** Read translations from file
* @param filename File to parse
* @return Array of translations
* @throws Exception on error (file not found, parse error)
*/
|
Read translations from file
|
getTranslations
|
{
"repo_name": "css-iter/cs-studio",
"path": "applications/alarm/alarm-plugins/org.csstudio.alarm.beast.annunciator/src/org/csstudio/alarm/beast/annunciator/model/TranslationFileReader.java",
"license": "epl-1.0",
"size": 3548
}
|
[
"java.io.BufferedReader",
"java.io.DataInputStream",
"java.io.FileInputStream",
"java.io.InputStreamReader",
"java.util.ArrayList",
"org.csstudio.utility.speech.Translation"
] |
import java.io.BufferedReader; import java.io.DataInputStream; import java.io.FileInputStream; import java.io.InputStreamReader; import java.util.ArrayList; import org.csstudio.utility.speech.Translation;
|
import java.io.*; import java.util.*; import org.csstudio.utility.speech.*;
|
[
"java.io",
"java.util",
"org.csstudio.utility"
] |
java.io; java.util; org.csstudio.utility;
| 823,317
|
protected String generateUniqueName() {
return generateUniqueName(name.getMethodName().replace('[', '-').replace("]", ""));
}
private static class DataFlowShell extends JLineShellComponent {
private final JLineShellComponent shell;
public DataFlowShell(JLineShellComponent shell) {
this.shell = shell;
}
|
String function() { return generateUniqueName(name.getMethodName().replace('[', '-').replace("]", "")); } private static class DataFlowShell extends JLineShellComponent { private final JLineShellComponent shell; public DataFlowShell(JLineShellComponent shell) { this.shell = shell; }
|
/**
* Return a unique random name for stream/task testing.
*
* @return unique random stream/task name
*/
|
Return a unique random name for stream/task testing
|
generateUniqueName
|
{
"repo_name": "jvalkeal/spring-cloud-data",
"path": "spring-cloud-dataflow-shell-core/src/test/java/org/springframework/cloud/dataflow/shell/AbstractShellIntegrationTest.java",
"license": "apache-2.0",
"size": 6888
}
|
[
"org.springframework.shell.core.JLineShellComponent"
] |
import org.springframework.shell.core.JLineShellComponent;
|
import org.springframework.shell.core.*;
|
[
"org.springframework.shell"
] |
org.springframework.shell;
| 1,437,293
|
private Hashtable<String, RegistryPackageDocument> transformSensorDescriptions(Collection<SirSearchResultElement> sensors) {
Hashtable<String, RegistryPackageDocument> transformedDocs = new Hashtable<>();
XmlObject description;
for (SirSearchResultElement sensorResultElem : sensors) {
log.debug("Transforming sensor description of sensor {}", sensorResultElem.getSensorId());
// get SensorML
SirXmlSensorDescription sensorDescr = (SirXmlSensorDescription) sensorResultElem.getSensorDescription();
description = sensorDescr.getDescription();
// having problems with XmlValueDisconnectedException, try transforming a deep copy
XmlObject copy = description.copy();
boolean isConform;
try {
isConform = acceptsDocument(copy);
}
catch (OwsExceptionReport | IOException e) {
log.error("Could not check if catalog accepts the given document!", e);
continue;
}
if (isConform) {
// transform SensorML
XmlObject ebrimDescription;
try {
if (copy instanceof SystemType) {
SystemType st = (SystemType) copy;
ebrimDescription = this.transformer.transform(st);
}
else if (copy instanceof SensorMLDocument) {
SensorMLDocument smlDoc = (SensorMLDocument) copy;
ebrimDescription = this.transformer.transform(smlDoc);
}
else {
throw new UnsupportedOperationException("Sensor description is of unsupported type (must be either SystemType or SensorMLDocument): "
+ copy.xmlText());
}
}
catch (XmlException | TransformerException | XmlValueDisconnectedException | IOException e) {
log.error("Exception: Could not transform sensor description: " + XmlTools.inspect(description)
+ "\n - TransformerException: " + e);
transformedDocs.put(sensorResultElem.getSensorId(), ITransformer.TRANSFORMATION_ERROR_OBJECT);
continue;
}
// convert transformed document to identifiable type
RegistryPackageDocument registryPackage;
try {
registryPackage = RegistryPackageDocument.Factory.parse(ebrimDescription.getDomNode());
}
catch (XmlException e) {
log.warn("Could not parse sensor description to IdentifiableType: "
+ XmlTools.inspect(ebrimDescription),
e);
transformedDocs.put(sensorResultElem.getSensorId(), ITransformer.TRANSFORMATION_ERROR_OBJECT);
continue;
}
if (SirConfigurator.getInstance().isValidateRequests()) {
// check if the transformed document for the outgoing request is valid and warn if not.
if (registryPackage.validate()) {
log.debug("Added new (valid!) transformed sensor description for sensor with id {}:\n{}",
sensorResultElem.getSensorId(),
XmlTools.inspect(ebrimDescription));
}
else {
String errors = XmlTools.validateAndIterateErrors(registryPackage);
log.warn("Transformed sensor description sensor with id " + sensorResultElem.getSensorId()
+ " IS NOT VALID and might not be accepted by the service.");
log.debug("\nErrors:\t{}\nebRIM:\t{}", errors, ebrimDescription.xmlText());
}
}
// add transformed document and sensor id
transformedDocs.put(sensorResultElem.getSensorId(), registryPackage);
log.info("Transformed sensor with id " + sensorResultElem.getSensorId()
+ " to RegistryPackage with id " + registryPackage.getIdentifiable().getId());
}
else {
log.warn("Could not transform sensor description of sensor " + sensorResultElem.getSensorId()
+ ". It is not conform with this catalog's required profile!");
transformedDocs.put(sensorResultElem.getSensorId(), ITransformer.TRANSFORMATION_ERROR_OBJECT);
}
} // for loop
return transformedDocs;
}
|
Hashtable<String, RegistryPackageDocument> function(Collection<SirSearchResultElement> sensors) { Hashtable<String, RegistryPackageDocument> transformedDocs = new Hashtable<>(); XmlObject description; for (SirSearchResultElement sensorResultElem : sensors) { log.debug(STR, sensorResultElem.getSensorId()); SirXmlSensorDescription sensorDescr = (SirXmlSensorDescription) sensorResultElem.getSensorDescription(); description = sensorDescr.getDescription(); XmlObject copy = description.copy(); boolean isConform; try { isConform = acceptsDocument(copy); } catch (OwsExceptionReport IOException e) { log.error(STR, e); continue; } if (isConform) { XmlObject ebrimDescription; try { if (copy instanceof SystemType) { SystemType st = (SystemType) copy; ebrimDescription = this.transformer.transform(st); } else if (copy instanceof SensorMLDocument) { SensorMLDocument smlDoc = (SensorMLDocument) copy; ebrimDescription = this.transformer.transform(smlDoc); } else { throw new UnsupportedOperationException(STR + copy.xmlText()); } } catch (XmlException TransformerException XmlValueDisconnectedException IOException e) { log.error(STR + XmlTools.inspect(description) + STR + e); transformedDocs.put(sensorResultElem.getSensorId(), ITransformer.TRANSFORMATION_ERROR_OBJECT); continue; } RegistryPackageDocument registryPackage; try { registryPackage = RegistryPackageDocument.Factory.parse(ebrimDescription.getDomNode()); } catch (XmlException e) { log.warn(STR + XmlTools.inspect(ebrimDescription), e); transformedDocs.put(sensorResultElem.getSensorId(), ITransformer.TRANSFORMATION_ERROR_OBJECT); continue; } if (SirConfigurator.getInstance().isValidateRequests()) { if (registryPackage.validate()) { log.debug(STR, sensorResultElem.getSensorId(), XmlTools.inspect(ebrimDescription)); } else { String errors = XmlTools.validateAndIterateErrors(registryPackage); log.warn(STR + sensorResultElem.getSensorId() + STR); log.debug(STR, errors, ebrimDescription.xmlText()); } } transformedDocs.put(sensorResultElem.getSensorId(), registryPackage); log.info(STR + sensorResultElem.getSensorId() + STR + registryPackage.getIdentifiable().getId()); } else { log.warn(STR + sensorResultElem.getSensorId() + STR); transformedDocs.put(sensorResultElem.getSensorId(), ITransformer.TRANSFORMATION_ERROR_OBJECT); } } return transformedDocs; }
|
/**
*
* Method tries to transform the given collection of sensor description documents. if there is a problem
* with one of the documents the returned list will contain the kvp <"sensorID", NULL>. This allows
* alerting the user to check the log for details.
*
* @param sensors
* @return
*/
|
Method tries to transform the given collection of sensor description documents. if there is a problem with one of the documents the returned list will contain the kvp . This allows alerting the user to check the log for details
|
transformSensorDescriptions
|
{
"repo_name": "52North/OpenSensorSearch",
"path": "service/src/main/java/org/n52/sir/catalog/csw/CswCatalog.java",
"license": "apache-2.0",
"size": 33984
}
|
[
"java.io.IOException",
"java.util.Collection",
"java.util.Hashtable",
"javax.xml.transform.TransformerException",
"net.opengis.sensorML.x101.SensorMLDocument",
"net.opengis.sensorML.x101.SystemType",
"org.apache.xmlbeans.XmlException",
"org.apache.xmlbeans.XmlObject",
"org.apache.xmlbeans.impl.values.XmlValueDisconnectedException",
"org.n52.oss.sir.api.SirSearchResultElement",
"org.n52.oss.sir.api.SirXmlSensorDescription",
"org.n52.oss.sir.ows.OwsExceptionReport",
"org.n52.oss.util.XmlTools",
"org.n52.sir.SirConfigurator",
"org.n52.sir.xml.ITransformer"
] |
import java.io.IOException; import java.util.Collection; import java.util.Hashtable; import javax.xml.transform.TransformerException; import net.opengis.sensorML.x101.SensorMLDocument; import net.opengis.sensorML.x101.SystemType; import org.apache.xmlbeans.XmlException; import org.apache.xmlbeans.XmlObject; import org.apache.xmlbeans.impl.values.XmlValueDisconnectedException; import org.n52.oss.sir.api.SirSearchResultElement; import org.n52.oss.sir.api.SirXmlSensorDescription; import org.n52.oss.sir.ows.OwsExceptionReport; import org.n52.oss.util.XmlTools; import org.n52.sir.SirConfigurator; import org.n52.sir.xml.ITransformer;
|
import java.io.*; import java.util.*; import javax.xml.transform.*; import net.opengis.*; import org.apache.xmlbeans.*; import org.apache.xmlbeans.impl.values.*; import org.n52.oss.sir.api.*; import org.n52.oss.sir.ows.*; import org.n52.oss.util.*; import org.n52.sir.*; import org.n52.sir.xml.*;
|
[
"java.io",
"java.util",
"javax.xml",
"net.opengis",
"org.apache.xmlbeans",
"org.n52.oss",
"org.n52.sir"
] |
java.io; java.util; javax.xml; net.opengis; org.apache.xmlbeans; org.n52.oss; org.n52.sir;
| 1,052,727
|
public ViewHolder setText(int viewId, String text){
TextView tv = getView(viewId);
tv.setText(text);
return this;
}
|
ViewHolder function(int viewId, String text){ TextView tv = getView(viewId); tv.setText(text); return this; }
|
/**
* set text
*/
|
set text
|
setText
|
{
"repo_name": "LSL-Git/ImageLabelerApp",
"path": "app/src/main/java/app/com/lsl/imagelabelerapp/lsl/utils/ViewHolder.java",
"license": "apache-2.0",
"size": 2244
}
|
[
"android.widget.TextView"
] |
import android.widget.TextView;
|
import android.widget.*;
|
[
"android.widget"
] |
android.widget;
| 2,323,840
|
public UniqueID getID();
|
UniqueID function();
|
/**
* Returns the unique id.
*
* @return The unique id of this active object.
*/
|
Returns the unique id
|
getID
|
{
"repo_name": "paraita/programming",
"path": "programming-core/src/main/java/org/objectweb/proactive/core/jmx/mbean/BodyWrapperMBean.java",
"license": "agpl-3.0",
"size": 2943
}
|
[
"org.objectweb.proactive.core.UniqueID"
] |
import org.objectweb.proactive.core.UniqueID;
|
import org.objectweb.proactive.core.*;
|
[
"org.objectweb.proactive"
] |
org.objectweb.proactive;
| 1,850,096
|
public CertificateOrderInner withProvisioningState(ProvisioningState provisioningState) {
this.provisioningState = provisioningState;
return this;
}
|
CertificateOrderInner function(ProvisioningState provisioningState) { this.provisioningState = provisioningState; return this; }
|
/**
* Set the provisioningState value.
*
* @param provisioningState the provisioningState value to set
* @return the CertificateOrderInner object itself.
*/
|
Set the provisioningState value
|
withProvisioningState
|
{
"repo_name": "herveyw/azure-sdk-for-java",
"path": "azure-mgmt-website/src/main/java/com/microsoft/azure/management/website/implementation/CertificateOrderInner.java",
"license": "mit",
"size": 11765
}
|
[
"com.microsoft.azure.management.website.ProvisioningState"
] |
import com.microsoft.azure.management.website.ProvisioningState;
|
import com.microsoft.azure.management.website.*;
|
[
"com.microsoft.azure"
] |
com.microsoft.azure;
| 401,651
|
logger.info("selectFeatures");
Configuration configuration = getConfiguration();
Dataframe[] data = Datasets.featureTransformationPCA(configuration);
Dataframe originalData = data[0];
Dataframe validationData = data[0].copy();
Dataframe expResult = data[1];
String storageName = this.getClass().getSimpleName();
PCA.TrainingParameters param = new PCA.TrainingParameters();
param.setMaxDimensions(null);
PCA instance = MLBuilder.create(param, configuration);
instance.fit_transform(originalData);
instance.save(storageName);
originalData.close();
instance.close();
instance = MLBuilder.load(PCA.class, storageName, configuration);
instance.transform(validationData);
assertEquals(validationData.size(), expResult.size());
Iterator<Record> itResult = validationData.iterator();
Iterator<Record> itExpectedResult = expResult.iterator();
while(itResult.hasNext()) {
Record r1 = itResult.next();
Record r2 = itExpectedResult.next();
for(Map.Entry<Object, Object> entry : r1.getX().entrySet()) {
Object feature = entry.getKey();
Double value = TypeInference.toDouble(entry.getValue());
assertEquals(TypeInference.toDouble(r2.getX().get(feature)), value, Constants.DOUBLE_ACCURACY_MEDIUM);
}
}
instance.delete();
validationData.close();
expResult.close();
}
|
logger.info(STR); Configuration configuration = getConfiguration(); Dataframe[] data = Datasets.featureTransformationPCA(configuration); Dataframe originalData = data[0]; Dataframe validationData = data[0].copy(); Dataframe expResult = data[1]; String storageName = this.getClass().getSimpleName(); PCA.TrainingParameters param = new PCA.TrainingParameters(); param.setMaxDimensions(null); PCA instance = MLBuilder.create(param, configuration); instance.fit_transform(originalData); instance.save(storageName); originalData.close(); instance.close(); instance = MLBuilder.load(PCA.class, storageName, configuration); instance.transform(validationData); assertEquals(validationData.size(), expResult.size()); Iterator<Record> itResult = validationData.iterator(); Iterator<Record> itExpectedResult = expResult.iterator(); while(itResult.hasNext()) { Record r1 = itResult.next(); Record r2 = itExpectedResult.next(); for(Map.Entry<Object, Object> entry : r1.getX().entrySet()) { Object feature = entry.getKey(); Double value = TypeInference.toDouble(entry.getValue()); assertEquals(TypeInference.toDouble(r2.getX().get(feature)), value, Constants.DOUBLE_ACCURACY_MEDIUM); } } instance.delete(); validationData.close(); expResult.close(); }
|
/**
* Test of selectFeatures method, of class PCA.
*/
|
Test of selectFeatures method, of class PCA
|
testSelectFeatures
|
{
"repo_name": "datumbox/datumbox-framework",
"path": "datumbox-framework-core/src/test/java/com/datumbox/framework/core/machinelearning/featureselection/PCATest.java",
"license": "apache-2.0",
"size": 3228
}
|
[
"com.datumbox.framework.common.Configuration",
"com.datumbox.framework.common.dataobjects.TypeInference",
"com.datumbox.framework.core.Datasets",
"com.datumbox.framework.core.common.dataobjects.Dataframe",
"com.datumbox.framework.core.common.dataobjects.Record",
"com.datumbox.framework.core.machinelearning.MLBuilder",
"com.datumbox.framework.tests.Constants",
"java.util.Iterator",
"java.util.Map",
"org.junit.Assert"
] |
import com.datumbox.framework.common.Configuration; import com.datumbox.framework.common.dataobjects.TypeInference; import com.datumbox.framework.core.Datasets; import com.datumbox.framework.core.common.dataobjects.Dataframe; import com.datumbox.framework.core.common.dataobjects.Record; import com.datumbox.framework.core.machinelearning.MLBuilder; import com.datumbox.framework.tests.Constants; import java.util.Iterator; import java.util.Map; import org.junit.Assert;
|
import com.datumbox.framework.common.*; import com.datumbox.framework.common.dataobjects.*; import com.datumbox.framework.core.*; import com.datumbox.framework.core.common.dataobjects.*; import com.datumbox.framework.core.machinelearning.*; import com.datumbox.framework.tests.*; import java.util.*; import org.junit.*;
|
[
"com.datumbox.framework",
"java.util",
"org.junit"
] |
com.datumbox.framework; java.util; org.junit;
| 2,113,702
|
@Override
public AppleBitcodeMode getAppleBitcodeMode() {
return appleBitcodeMode;
}
|
AppleBitcodeMode function() { return appleBitcodeMode; }
|
/**
* Returns the bitcode mode to use for compilation.
*
* <p>Users can control bitcode mode using the {@code apple_bitcode} build flag, but bitcode will
* be disabled for all simulator architectures regardless of this flag.
*/
|
Returns the bitcode mode to use for compilation. Users can control bitcode mode using the apple_bitcode build flag, but bitcode will be disabled for all simulator architectures regardless of this flag
|
getAppleBitcodeMode
|
{
"repo_name": "twitter-forks/bazel",
"path": "src/main/java/com/google/devtools/build/lib/rules/cpp/CppConfiguration.java",
"license": "apache-2.0",
"size": 29964
}
|
[
"com.google.devtools.build.lib.rules.apple.AppleCommandLineOptions"
] |
import com.google.devtools.build.lib.rules.apple.AppleCommandLineOptions;
|
import com.google.devtools.build.lib.rules.apple.*;
|
[
"com.google.devtools"
] |
com.google.devtools;
| 412,506
|
private void createMapAnnotationandLinkToProject(long projectId)
throws Exception
{
List<NamedValue> result = new ArrayList<NamedValue>();
result.add(new NamedValue("mitomycin-A", "20mM"));
result.add(new NamedValue("PBS", "10mM"));
result.add(new NamedValue("incubation", "5min"));
result.add(new NamedValue("temperature", "37"));
result.add(new NamedValue("Organism", "Homo sapiens"));
MapAnnotationData data = new MapAnnotationData();
data.setContent(result);
data.setDescription("Training Example");
//Use the following namespace if you want the annotation to be editable
//in the webclient and insight
data.setNameSpace(MapAnnotationData.NS_CLIENT_CREATED);
DataManagerFacility fac = gateway.getFacility(DataManagerFacility.class);
fac.attachAnnotation(ctx, data, new ProjectData(new ProjectI(projectId, false)));
}
// Create file annotation
// ======================
|
void function(long projectId) throws Exception { List<NamedValue> result = new ArrayList<NamedValue>(); result.add(new NamedValue(STR, "20mM")); result.add(new NamedValue("PBS", "10mM")); result.add(new NamedValue(STR, "5min")); result.add(new NamedValue(STR, "37")); result.add(new NamedValue(STR, STR)); MapAnnotationData data = new MapAnnotationData(); data.setContent(result); data.setDescription(STR); data.setNameSpace(MapAnnotationData.NS_CLIENT_CREATED); DataManagerFacility fac = gateway.getFacility(DataManagerFacility.class); fac.attachAnnotation(ctx, data, new ProjectData(new ProjectI(projectId, false))); }
|
/**
* Creates a map annotation and links to the specified project.
* @param projectId The omero project identifier
* @throws Exception
*/
|
Creates a map annotation and links to the specified project
|
createMapAnnotationandLinkToProject
|
{
"repo_name": "knabar/openmicroscopy",
"path": "examples/Training/java/src/training/WriteData.java",
"license": "gpl-2.0",
"size": 15632
}
|
[
"java.util.ArrayList",
"java.util.List"
] |
import java.util.ArrayList; import java.util.List;
|
import java.util.*;
|
[
"java.util"
] |
java.util;
| 2,229,375
|
@XmlAttribute
public void setErrorHandlerRef(String errorHandlerRef) {
this.errorHandlerRef = errorHandlerRef;
// we use an specific error handler ref (from Spring DSL) then wrap that
// with a error handler build ref so Camel knows its not just the default one
setErrorHandlerBuilder(new ErrorHandlerBuilderRef(errorHandlerRef));
}
|
void function(String errorHandlerRef) { this.errorHandlerRef = errorHandlerRef; setErrorHandlerBuilder(new ErrorHandlerBuilderRef(errorHandlerRef)); }
|
/**
* Sets the bean ref name of the error handler builder to use on this route
*/
|
Sets the bean ref name of the error handler builder to use on this route
|
setErrorHandlerRef
|
{
"repo_name": "jonmcewen/camel",
"path": "camel-core/src/main/java/org/apache/camel/model/RouteDefinition.java",
"license": "apache-2.0",
"size": 44503
}
|
[
"org.apache.camel.builder.ErrorHandlerBuilderRef"
] |
import org.apache.camel.builder.ErrorHandlerBuilderRef;
|
import org.apache.camel.builder.*;
|
[
"org.apache.camel"
] |
org.apache.camel;
| 1,538,193
|
public NestedSet<LibraryToLink> getLibraries() {
return this.libraries;
}
|
NestedSet<LibraryToLink> function() { return this.libraries; }
|
/**
* Returns libraries that are to be inputs to the linker.
*/
|
Returns libraries that are to be inputs to the linker
|
getLibraries
|
{
"repo_name": "juhalindfors/bazel-patches",
"path": "src/main/java/com/google/devtools/build/lib/rules/cpp/CppLinkAction.java",
"license": "apache-2.0",
"size": 25946
}
|
[
"com.google.devtools.build.lib.collect.nestedset.NestedSet",
"com.google.devtools.build.lib.rules.cpp.LinkerInputs"
] |
import com.google.devtools.build.lib.collect.nestedset.NestedSet; import com.google.devtools.build.lib.rules.cpp.LinkerInputs;
|
import com.google.devtools.build.lib.collect.nestedset.*; import com.google.devtools.build.lib.rules.cpp.*;
|
[
"com.google.devtools"
] |
com.google.devtools;
| 1,189,133
|
List<GwasDTO> mappedList = new ArrayList<>();
for (GwasDTO gwasMapping : gwasMappings) {
if (gwasMapping.getGwasMgiGeneSymbol().equals(mgiGeneSymbol)) {
mappedList.add(gwasMapping);
}
return mappedList;
}
return null;
}
|
List<GwasDTO> mappedList = new ArrayList<>(); for (GwasDTO gwasMapping : gwasMappings) { if (gwasMapping.getGwasMgiGeneSymbol().equals(mgiGeneSymbol)) { mappedList.add(gwasMapping); } return mappedList; } return null; }
|
/**
* Fetch all gwas mapping rows filtered by mgi gene sysmbol.
*
* @return all gwas mapping rows filtered by mgi gene symbol
* @throws SQLException
*/
|
Fetch all gwas mapping rows filtered by mgi gene sysmbol
|
getGwasMappingByGeneSymbol
|
{
"repo_name": "mpi2/PhenotypeArchive",
"path": "src/main/java/uk/ac/ebi/phenotype/dao/GwasDAO.java",
"license": "apache-2.0",
"size": 14613
}
|
[
"java.util.ArrayList",
"java.util.List",
"uk.ac.ebi.phenotype.service.dto.GwasDTO"
] |
import java.util.ArrayList; import java.util.List; import uk.ac.ebi.phenotype.service.dto.GwasDTO;
|
import java.util.*; import uk.ac.ebi.phenotype.service.dto.*;
|
[
"java.util",
"uk.ac.ebi"
] |
java.util; uk.ac.ebi;
| 320,457
|
@Override
public List<IItemPropertyDescriptor> getPropertyDescriptors(Object object)
{
if (itemPropertyDescriptors == null)
{
super.getPropertyDescriptors(object);
}
return itemPropertyDescriptors;
}
|
List<IItemPropertyDescriptor> function(Object object) { if (itemPropertyDescriptors == null) { super.getPropertyDescriptors(object); } return itemPropertyDescriptors; }
|
/**
* This returns the property descriptors for the adapted class.
* <!-- begin-user-doc -->
* <!-- end-user-doc -->
* @generated
*/
|
This returns the property descriptors for the adapted class.
|
getPropertyDescriptors
|
{
"repo_name": "peterkir/org.eclipse.oomph",
"path": "plugins/org.eclipse.oomph.setup.edit/src/org/eclipse/oomph/setup/provider/SetupTaskContainerItemProvider.java",
"license": "epl-1.0",
"size": 8001
}
|
[
"java.util.List",
"org.eclipse.emf.edit.provider.IItemPropertyDescriptor"
] |
import java.util.List; import org.eclipse.emf.edit.provider.IItemPropertyDescriptor;
|
import java.util.*; import org.eclipse.emf.edit.provider.*;
|
[
"java.util",
"org.eclipse.emf"
] |
java.util; org.eclipse.emf;
| 1,371,775
|
private static void addExtraReads(ModuleLayer bootLayer) {
// decode the command line options
Map<String, List<String>> map = decode("jdk.module.addreads.");
if (map.isEmpty())
return;
for (Map.Entry<String, List<String>> e : map.entrySet()) {
// the key is $MODULE
String mn = e.getKey();
Optional<Module> om = bootLayer.findModule(mn);
if (!om.isPresent()) {
warnUnknownModule(ADD_READS, mn);
continue;
}
Module m = om.get();
// the value is the set of other modules (by name)
for (String name : e.getValue()) {
if (ALL_UNNAMED.equals(name)) {
Modules.addReadsAllUnnamed(m);
} else {
om = bootLayer.findModule(name);
if (om.isPresent()) {
Modules.addReads(m, om.get());
} else {
warnUnknownModule(ADD_READS, name);
}
}
}
}
}
|
static void function(ModuleLayer bootLayer) { Map<String, List<String>> map = decode(STR); if (map.isEmpty()) return; for (Map.Entry<String, List<String>> e : map.entrySet()) { String mn = e.getKey(); Optional<Module> om = bootLayer.findModule(mn); if (!om.isPresent()) { warnUnknownModule(ADD_READS, mn); continue; } Module m = om.get(); for (String name : e.getValue()) { if (ALL_UNNAMED.equals(name)) { Modules.addReadsAllUnnamed(m); } else { om = bootLayer.findModule(name); if (om.isPresent()) { Modules.addReads(m, om.get()); } else { warnUnknownModule(ADD_READS, name); } } } } }
|
/**
* Process the --add-reads options to add any additional read edges that
* are specified on the command-line.
*/
|
Process the --add-reads options to add any additional read edges that are specified on the command-line
|
addExtraReads
|
{
"repo_name": "md-5/jdk10",
"path": "src/java.base/share/classes/jdk/internal/module/ModuleBootstrap.java",
"license": "gpl-2.0",
"size": 41334
}
|
[
"java.util.List",
"java.util.Map",
"java.util.Optional"
] |
import java.util.List; import java.util.Map; import java.util.Optional;
|
import java.util.*;
|
[
"java.util"
] |
java.util;
| 1,689,997
|
protected void addOldViewNamePropertyDescriptor(Object object) {
itemPropertyDescriptors.add
(createItemPropertyDescriptor
(((ComposeableAdapterFactory)adapterFactory).getRootAdapterFactory(),
getResourceLocator(),
getString("_UI_RenameViewType_oldViewName_feature"),
getString("_UI_PropertyDescriptor_description", "_UI_RenameViewType_oldViewName_feature", "_UI_RenameViewType_type"),
DbchangelogPackage.eINSTANCE.getRenameViewType_OldViewName(),
true,
false,
false,
ItemPropertyDescriptor.GENERIC_VALUE_IMAGE,
null,
null));
}
|
void function(Object object) { itemPropertyDescriptors.add (createItemPropertyDescriptor (((ComposeableAdapterFactory)adapterFactory).getRootAdapterFactory(), getResourceLocator(), getString(STR), getString(STR, STR, STR), DbchangelogPackage.eINSTANCE.getRenameViewType_OldViewName(), true, false, false, ItemPropertyDescriptor.GENERIC_VALUE_IMAGE, null, null)); }
|
/**
* This adds a property descriptor for the Old View Name feature.
* <!-- begin-user-doc -->
* <!-- end-user-doc -->
* @generated
*/
|
This adds a property descriptor for the Old View Name feature.
|
addOldViewNamePropertyDescriptor
|
{
"repo_name": "Treehopper/EclipseAugments",
"path": "liquibase-editor/eu.hohenegger.xsd.liquibase.ui/src-gen/org/liquibase/xml/ns/dbchangelog/provider/RenameViewTypeItemProvider.java",
"license": "epl-1.0",
"size": 8770
}
|
[
"org.eclipse.emf.edit.provider.ComposeableAdapterFactory",
"org.eclipse.emf.edit.provider.ItemPropertyDescriptor",
"org.liquibase.xml.ns.dbchangelog.DbchangelogPackage"
] |
import org.eclipse.emf.edit.provider.ComposeableAdapterFactory; import org.eclipse.emf.edit.provider.ItemPropertyDescriptor; import org.liquibase.xml.ns.dbchangelog.DbchangelogPackage;
|
import org.eclipse.emf.edit.provider.*; import org.liquibase.xml.ns.dbchangelog.*;
|
[
"org.eclipse.emf",
"org.liquibase.xml"
] |
org.eclipse.emf; org.liquibase.xml;
| 594,319
|
public void addMemberHeader(ExecutableElement member, Content methodsContentTree) {
methodsContentTree.addContent(getHead(member));
methodsContentTree.addContent(getSignature(member));
}
|
void function(ExecutableElement member, Content methodsContentTree) { methodsContentTree.addContent(getHead(member)); methodsContentTree.addContent(getSignature(member)); }
|
/**
* Add the member header.
*
* @param member the method document to be listed
* @param methodsContentTree the content tree to which the member header will be added
*/
|
Add the member header
|
addMemberHeader
|
{
"repo_name": "FauxFaux/jdk9-langtools",
"path": "src/jdk.javadoc/share/classes/jdk/javadoc/internal/doclets/formats/html/HtmlSerialMethodWriter.java",
"license": "gpl-2.0",
"size": 6699
}
|
[
"javax.lang.model.element.ExecutableElement"
] |
import javax.lang.model.element.ExecutableElement;
|
import javax.lang.model.element.*;
|
[
"javax.lang"
] |
javax.lang;
| 1,120,744
|
Set<Class<?>> findImplementations(Class<?> parent, String... packageNames);
|
Set<Class<?>> findImplementations(Class<?> parent, String... packageNames);
|
/**
* Attempts to discover classes that are assignable to the type provided. In
* the case that an interface is provided this method will collect
* implementations. In the case of a non-interface class, subclasses will be
* collected.
*
* @param parent the class of interface to find subclasses or implementations of
* @param packageNames one or more package names to scan (including subpackages) for classes
* @return the classes found, returns an empty set if none found
*/
|
Attempts to discover classes that are assignable to the type provided. In the case that an interface is provided this method will collect implementations. In the case of a non-interface class, subclasses will be collected
|
findImplementations
|
{
"repo_name": "Fabryprog/camel",
"path": "core/camel-api/src/main/java/org/apache/camel/spi/PackageScanClassResolver.java",
"license": "apache-2.0",
"size": 3992
}
|
[
"java.util.Set"
] |
import java.util.Set;
|
import java.util.*;
|
[
"java.util"
] |
java.util;
| 1,057,267
|
public List<Integer> probReachStrategy(STPG stpg, int state, BitSet target, boolean min1, boolean min2, double lastSoln[]) throws PrismException
{
double val = stpg.mvMultMinMaxSingle(state, lastSoln, min1, min2);
return stpg.mvMultMinMaxSingleChoices(state, lastSoln, min1, min2, val);
}
|
List<Integer> function(STPG stpg, int state, BitSet target, boolean min1, boolean min2, double lastSoln[]) throws PrismException { double val = stpg.mvMultMinMaxSingle(state, lastSoln, min1, min2); return stpg.mvMultMinMaxSingleChoices(state, lastSoln, min1, min2, val); }
|
/**
* Construct strategy information for min/max reachability probabilities.
* (More precisely, list of indices of player 1 choices resulting in min/max.)
* (Note: indices are guaranteed to be sorted in ascending order.)
* @param stpg The STPG
* @param state The state to generate strategy info for
* @param target The set of target states to reach
* @param min1 Min or max probabilities for player 1 (true=min, false=max)
* @param min2 Min or max probabilities for player 2 (true=min, false=max)
* @param lastSoln Vector of probabilities from which to recompute in one iteration
*/
|
Construct strategy information for min/max reachability probabilities. (More precisely, list of indices of player 1 choices resulting in min/max.) (Note: indices are guaranteed to be sorted in ascending order.)
|
probReachStrategy
|
{
"repo_name": "nicodelpiano/prism",
"path": "src/explicit/STPGModelChecker.java",
"license": "gpl-2.0",
"size": 32072
}
|
[
"java.util.BitSet",
"java.util.List"
] |
import java.util.BitSet; import java.util.List;
|
import java.util.*;
|
[
"java.util"
] |
java.util;
| 821,121
|
private Node tryFoldHook(Node n) {
Preconditions.checkState(n.isHook(), n);
Node parent = n.getParent();
Preconditions.checkNotNull(parent);
Node cond = n.getFirstChild();
Node thenBody = cond.getNext();
Node elseBody = thenBody.getNext();
TernaryValue condValue = NodeUtil.getImpureBooleanValue(cond);
if (condValue == TernaryValue.UNKNOWN) {
// If the result nodes are equivalent, then one of the nodes can be
// removed and it doesn't matter which.
if (!areNodesEqualForInlining(thenBody, elseBody)) {
return n; // We can't remove branches otherwise!
}
}
// Transform "(a = 2) ? x =2 : y" into "a=2,x=2"
Node branchToKeep = condValue.toBoolean(true) ? thenBody : elseBody;
Node replacement;
boolean condHasSideEffects = mayHaveSideEffects(cond);
// Must detach after checking for side effects, to ensure that the parents
// of nodes are set correctly.
n.detachChildren();
if (condHasSideEffects) {
replacement = IR.comma(cond, branchToKeep).srcref(n);
} else {
replacement = branchToKeep;
}
parent.replaceChild(n, replacement);
reportCodeChange();
return replacement;
}
|
Node function(Node n) { Preconditions.checkState(n.isHook(), n); Node parent = n.getParent(); Preconditions.checkNotNull(parent); Node cond = n.getFirstChild(); Node thenBody = cond.getNext(); Node elseBody = thenBody.getNext(); TernaryValue condValue = NodeUtil.getImpureBooleanValue(cond); if (condValue == TernaryValue.UNKNOWN) { if (!areNodesEqualForInlining(thenBody, elseBody)) { return n; } } Node branchToKeep = condValue.toBoolean(true) ? thenBody : elseBody; Node replacement; boolean condHasSideEffects = mayHaveSideEffects(cond); n.detachChildren(); if (condHasSideEffects) { replacement = IR.comma(cond, branchToKeep).srcref(n); } else { replacement = branchToKeep; } parent.replaceChild(n, replacement); reportCodeChange(); return replacement; }
|
/**
* Try folding HOOK (?:) if the condition results of the condition is known.
* @return the replacement node, if changed, or the original if not
*/
|
Try folding HOOK (?:) if the condition results of the condition is known
|
tryFoldHook
|
{
"repo_name": "redforks/closure-compiler",
"path": "src/com/google/javascript/jscomp/PeepholeRemoveDeadCode.java",
"license": "apache-2.0",
"size": 32888
}
|
[
"com.google.common.base.Preconditions",
"com.google.javascript.rhino.IR",
"com.google.javascript.rhino.Node",
"com.google.javascript.rhino.jstype.TernaryValue"
] |
import com.google.common.base.Preconditions; import com.google.javascript.rhino.IR; import com.google.javascript.rhino.Node; import com.google.javascript.rhino.jstype.TernaryValue;
|
import com.google.common.base.*; import com.google.javascript.rhino.*; import com.google.javascript.rhino.jstype.*;
|
[
"com.google.common",
"com.google.javascript"
] |
com.google.common; com.google.javascript;
| 2,793,446
|
@Test
public void testFormatWithNonInteractive() throws IOException {
// we check for a non empty dir, so create a child path
File data = new File(hdfsDir, "file");
if (!data.mkdirs()) {
fail("Failed to create dir " + data.getPath());
}
String[] argv = { "-format", "-nonInteractive" };
try {
NameNode.createNameNode(argv, config);
fail("createNameNode() did not call System.exit()");
} catch (ExitException e) {
assertEquals("Format should have been aborted with exit code 1", 1,
e.status);
}
// check if the version file does not exists.
File version = new File(hdfsDir, "current/VERSION");
assertFalse("Check version should not exist", version.exists());
}
|
void function() throws IOException { File data = new File(hdfsDir, "file"); if (!data.mkdirs()) { fail(STR + data.getPath()); } String[] argv = { STR, STR }; try { NameNode.createNameNode(argv, config); fail(STR); } catch (ExitException e) { assertEquals(STR, 1, e.status); } File version = new File(hdfsDir, STR); assertFalse(STR, version.exists()); }
|
/**
* Test namenode format with -format -nonInteractive options when a non empty
* name directory exists. Format should not succeed.
*
* @throws IOException
*/
|
Test namenode format with -format -nonInteractive options when a non empty name directory exists. Format should not succeed
|
testFormatWithNonInteractive
|
{
"repo_name": "NJUJYB/disYarn",
"path": "hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestClusterId.java",
"license": "apache-2.0",
"size": 14538
}
|
[
"java.io.File",
"java.io.IOException",
"org.apache.hadoop.util.ExitUtil",
"org.junit.Assert"
] |
import java.io.File; import java.io.IOException; import org.apache.hadoop.util.ExitUtil; import org.junit.Assert;
|
import java.io.*; import org.apache.hadoop.util.*; import org.junit.*;
|
[
"java.io",
"org.apache.hadoop",
"org.junit"
] |
java.io; org.apache.hadoop; org.junit;
| 2,761,366
|
public void optimize( Graphics2D g )
throws Exception {
for ( Optimizer optimizer : optimizers ) {
optimizer.optimize( g );
}
}
|
void function( Graphics2D g ) throws Exception { for ( Optimizer optimizer : optimizers ) { optimizer.optimize( g ); } }
|
/**
* Performs the optimization for all contained {@link Optimizer} instances. Calls
* {@link Optimizer#optimize(Graphics2D)} for all contained {@link Optimizer} instances. subsequently.
*
* @param g
*/
|
Performs the optimization for all contained <code>Optimizer</code> instances. Calls <code>Optimizer#optimize(Graphics2D)</code> for all contained <code>Optimizer</code> instances. subsequently
|
optimize
|
{
"repo_name": "lat-lon/deegree2-base",
"path": "deegree2-core/src/main/java/org/deegree/graphics/optimizers/OptimizerChain.java",
"license": "lgpl-2.1",
"size": 3067
}
|
[
"java.awt.Graphics2D"
] |
import java.awt.Graphics2D;
|
import java.awt.*;
|
[
"java.awt"
] |
java.awt;
| 2,759,814
|
private ProcessRequestScheduleEntityBean getProcessRequestScheduleEntityBean(
CReqInstructionLog reqInstructionLog, Integer requestId,
String emailsGroup, String processClassName, String currentUser,
Timestamp currentDate) throws CommDatabaseException {
ProcessRequestScheduleEntityBean scheduleBean = null;
String instructionName = null;
String instructionValue = null;
SimpleDateFormat sdf = new SimpleDateFormat("MM/dd/yyyy hh:mm:ss");
List<InstructionParameters> list = reqInstructionLog
.getInstructionParametersList();
KEEP_ALIVE keepAlive = KEEP_ALIVE.NO;
if(list != null) {
for (InstructionParameters instructionParameters : list) {
instructionName = instructionParameters.getName();
instructionValue = instructionParameters.getValue();
if (CommConstants.SCHEDULE_INSTRCUTION_PARAMS.BATCH_NAME.name()
.equals(instructionName)) {
scheduledBatchName = instructionValue;
} else if (CommConstants.SCHEDULE_INSTRCUTION_PARAMS.FREQUENCY.name()
.equals(instructionName)) {
if (scheduleBean == null) {
scheduleBean = new ProcessRequestScheduleEntityBean();
}
scheduleBean.setFreqType(instructionValue.toUpperCase());
} else if (CommConstants.SCHEDULE_INSTRCUTION_PARAMS.RECUR_EVERY
.name().equals(instructionName)) {
if (scheduleBean == null) {
scheduleBean = new ProcessRequestScheduleEntityBean();
}
scheduleBean.setRecur(Integer.parseInt(instructionValue));
// } else if (CommConstants.SCHEDULE_INSTRCUTION_PARAMS.BATCH_RUN_DATE
// .name().equals(instructionName)) {
//
// Date parsedDate;
// try {
// parsedDate = dateFormat2.parse(instructionValue);
// if (scheduleBean == null) {
// scheduleBean = new ProcessRequestScheduleEntityBean();
// }
// scheduleBean
// .setStartDt(new Timestamp(parsedDate.getTime()));
// } catch (ParseException e) {
// // Dummy
// e.printStackTrace();
// }
}else if (CommConstants.SCHEDULE_INSTRCUTION_PARAMS.SCHEDULE_DATE
.name().equals(instructionName)) {
Date parsedDate;
try {
parsedDate = sdf.parse(instructionValue);
scheduleTime = new Timestamp(parsedDate.getTime());
} catch (ParseException e) {
// Dummy
}
}else if (CommConstants.SCHEDULE_INSTRCUTION_PARAMS.WEEK_DAY
.name().equals(instructionName)) {
if (scheduleBean == null) {
scheduleBean = new ProcessRequestScheduleEntityBean();
}
scheduleBean.setOnWeekDay(instructionValue);
} else if (CommConstants.SCHEDULE_INSTRCUTION_PARAMS.END_ON_DATE
.name().equals(instructionName)) {
Date parsedDate;
try {
parsedDate = sdf.parse(instructionValue);
if (scheduleBean == null) {
scheduleBean = new ProcessRequestScheduleEntityBean();
}
scheduleBean
.setEndDt(new Timestamp(parsedDate.getTime()));
} catch (ParseException e) {
// Dummy
}
} else if (CommConstants.SCHEDULE_INSTRCUTION_PARAMS.END_ON_OCCURRENCE
.name().equals(instructionName)) {
if (scheduleBean == null) {
scheduleBean = new ProcessRequestScheduleEntityBean();
}
scheduleBean.setEndOccur(Long.parseLong(instructionValue));
} else if (CommConstants.SCHEDULE_INSTRCUTION_PARAMS.SKIP_FLAG
.name().equals(instructionName)) {
if (scheduleBean == null) {
scheduleBean = new ProcessRequestScheduleEntityBean();
}
scheduleBean.setSkipFlag(instructionValue);
} else if (CommConstants.SCHEDULE_INSTRCUTION_PARAMS.KEEP_ALIVE.name().equals(instructionName)) {
keepAlive = KEEP_ALIVE.resolve(instructionValue);
}
}
}
if (scheduleBean != null) {
scheduleBean.setSchId(requestId);
scheduleBean.setSchStat(ISchedule.SCHEDULE_STATUS.ACTIVE.getID());
scheduleBean.setStartDt(scheduleTime);
scheduleBean.setUserId(currentUser);
scheduleBean.setEntryDt(currentDate);
scheduleBean.setReqStat(IProcessRequest.REQUEST_STATUS.QUEUED.getID());
scheduleBean.setOccurCounter(1);
scheduleBean.setProcessClassNm(processClassName);
scheduleBean.setFutureSchedulingOnly(ISchedule.FUTURE_SCHEDULING_ONLY);
scheduleBean.setFixedDate("N");
scheduleBean.setEmailIds(emailsGroup);
scheduleBean.setWeekdayCheckFlag("N");
scheduleBean.setKeepAlive(keepAlive.getID());
}
return scheduleBean;
}
|
ProcessRequestScheduleEntityBean function( CReqInstructionLog reqInstructionLog, Integer requestId, String emailsGroup, String processClassName, String currentUser, Timestamp currentDate) throws CommDatabaseException { ProcessRequestScheduleEntityBean scheduleBean = null; String instructionName = null; String instructionValue = null; SimpleDateFormat sdf = new SimpleDateFormat(STR); List<InstructionParameters> list = reqInstructionLog .getInstructionParametersList(); KEEP_ALIVE keepAlive = KEEP_ALIVE.NO; if(list != null) { for (InstructionParameters instructionParameters : list) { instructionName = instructionParameters.getName(); instructionValue = instructionParameters.getValue(); if (CommConstants.SCHEDULE_INSTRCUTION_PARAMS.BATCH_NAME.name() .equals(instructionName)) { scheduledBatchName = instructionValue; } else if (CommConstants.SCHEDULE_INSTRCUTION_PARAMS.FREQUENCY.name() .equals(instructionName)) { if (scheduleBean == null) { scheduleBean = new ProcessRequestScheduleEntityBean(); } scheduleBean.setFreqType(instructionValue.toUpperCase()); } else if (CommConstants.SCHEDULE_INSTRCUTION_PARAMS.RECUR_EVERY .name().equals(instructionName)) { if (scheduleBean == null) { scheduleBean = new ProcessRequestScheduleEntityBean(); } scheduleBean.setRecur(Integer.parseInt(instructionValue)); }else if (CommConstants.SCHEDULE_INSTRCUTION_PARAMS.SCHEDULE_DATE .name().equals(instructionName)) { Date parsedDate; try { parsedDate = sdf.parse(instructionValue); scheduleTime = new Timestamp(parsedDate.getTime()); } catch (ParseException e) { } }else if (CommConstants.SCHEDULE_INSTRCUTION_PARAMS.WEEK_DAY .name().equals(instructionName)) { if (scheduleBean == null) { scheduleBean = new ProcessRequestScheduleEntityBean(); } scheduleBean.setOnWeekDay(instructionValue); } else if (CommConstants.SCHEDULE_INSTRCUTION_PARAMS.END_ON_DATE .name().equals(instructionName)) { Date parsedDate; try { parsedDate = sdf.parse(instructionValue); if (scheduleBean == null) { scheduleBean = new ProcessRequestScheduleEntityBean(); } scheduleBean .setEndDt(new Timestamp(parsedDate.getTime())); } catch (ParseException e) { } } else if (CommConstants.SCHEDULE_INSTRCUTION_PARAMS.END_ON_OCCURRENCE .name().equals(instructionName)) { if (scheduleBean == null) { scheduleBean = new ProcessRequestScheduleEntityBean(); } scheduleBean.setEndOccur(Long.parseLong(instructionValue)); } else if (CommConstants.SCHEDULE_INSTRCUTION_PARAMS.SKIP_FLAG .name().equals(instructionName)) { if (scheduleBean == null) { scheduleBean = new ProcessRequestScheduleEntityBean(); } scheduleBean.setSkipFlag(instructionValue); } else if (CommConstants.SCHEDULE_INSTRCUTION_PARAMS.KEEP_ALIVE.name().equals(instructionName)) { keepAlive = KEEP_ALIVE.resolve(instructionValue); } } } if (scheduleBean != null) { scheduleBean.setSchId(requestId); scheduleBean.setSchStat(ISchedule.SCHEDULE_STATUS.ACTIVE.getID()); scheduleBean.setStartDt(scheduleTime); scheduleBean.setUserId(currentUser); scheduleBean.setEntryDt(currentDate); scheduleBean.setReqStat(IProcessRequest.REQUEST_STATUS.QUEUED.getID()); scheduleBean.setOccurCounter(1); scheduleBean.setProcessClassNm(processClassName); scheduleBean.setFutureSchedulingOnly(ISchedule.FUTURE_SCHEDULING_ONLY); scheduleBean.setFixedDate("N"); scheduleBean.setEmailIds(emailsGroup); scheduleBean.setWeekdayCheckFlag("N"); scheduleBean.setKeepAlive(keepAlive.getID()); } return scheduleBean; }
|
/**
* Creates (Inserts) records in the process_request_schedule table.
*
* @param reqInstructionLog
* The instruction log to be inserted in INSTRUCTION_LOG of CORE @link
* CReqInstructionLog
*
* @param requestId
* Instruction request id.
*
* @param currentUser
* The current application user
*
* @param currentDate
* The system date
*
* @param processClassName
* The process class name
*
* @param emailsGroup
* The email group to which the notification regarding batch
* process will be sent
*
* @throws CommDatabaseException
* Any database related I/O exception
*/
|
Creates (Inserts) records in the process_request_schedule table
|
getProcessRequestScheduleEntityBean
|
{
"repo_name": "MastekLtd/JBEAM",
"path": "supporting_libraries/jbeam-plugins/jbeam-core-comm-oracle/src/main/java/com/stgmastek/core/comm/server/dao/impl/BatchDAO.java",
"license": "lgpl-3.0",
"size": 43501
}
|
[
"com.stgmastek.core.comm.exception.CommDatabaseException",
"com.stgmastek.core.comm.server.vo.CReqInstructionLog",
"com.stgmastek.core.comm.server.vo.InstructionParameters",
"com.stgmastek.core.comm.util.CommConstants",
"java.sql.Timestamp",
"java.text.ParseException",
"java.text.SimpleDateFormat",
"java.util.Date",
"java.util.List"
] |
import com.stgmastek.core.comm.exception.CommDatabaseException; import com.stgmastek.core.comm.server.vo.CReqInstructionLog; import com.stgmastek.core.comm.server.vo.InstructionParameters; import com.stgmastek.core.comm.util.CommConstants; import java.sql.Timestamp; import java.text.ParseException; import java.text.SimpleDateFormat; import java.util.Date; import java.util.List;
|
import com.stgmastek.core.comm.exception.*; import com.stgmastek.core.comm.server.vo.*; import com.stgmastek.core.comm.util.*; import java.sql.*; import java.text.*; import java.util.*;
|
[
"com.stgmastek.core",
"java.sql",
"java.text",
"java.util"
] |
com.stgmastek.core; java.sql; java.text; java.util;
| 1,181,711
|
public static String netconfGet(NetconfSession session, String filter) {
String reply;
try {
reply = session.get(filter, null);
} catch (NetconfException e) {
throw new IllegalStateException(new NetconfException("Failed to retrieve configuration.", e));
}
return reply;
}
|
static String function(NetconfSession session, String filter) { String reply; try { reply = session.get(filter, null); } catch (NetconfException e) { throw new IllegalStateException(new NetconfException(STR, e)); } return reply; }
|
/**
* Retrieves session reply information for get operation.
*
* @param session explicit NETCONF session
* @param filter the filter string of xml content
* @return the reply string
*/
|
Retrieves session reply information for get operation
|
netconfGet
|
{
"repo_name": "opennetworkinglab/onos",
"path": "drivers/polatis/netconf/src/main/java/org/onosproject/drivers/polatis/netconf/PolatisNetconfUtility.java",
"license": "apache-2.0",
"size": 11359
}
|
[
"org.onosproject.netconf.NetconfException",
"org.onosproject.netconf.NetconfSession"
] |
import org.onosproject.netconf.NetconfException; import org.onosproject.netconf.NetconfSession;
|
import org.onosproject.netconf.*;
|
[
"org.onosproject.netconf"
] |
org.onosproject.netconf;
| 1,417,798
|
public static <T> double RR(List<T> rankedList, List<T> groundTruth) {
for (int i = 0, n = rankedList.size(); i < n; i++) {
T item = rankedList.get(i);
if (groundTruth.contains(item))
return 1 / (i + 1.0);
}
return 0;
}
|
static <T> double function(List<T> rankedList, List<T> groundTruth) { for (int i = 0, n = rankedList.size(); i < n; i++) { T item = rankedList.get(i); if (groundTruth.contains(item)) return 1 / (i + 1.0); } return 0; }
|
/**
* Compute the reciprocal rank of a list of ranked items
*
* @param <T>
*
* @param rankedList
* a list of ranked item IDs, the highest-ranking item first
* @param groundTruth
* a collection of positive/correct item IDs
* @return the mean reciprocal rank for the given data<
*/
|
Compute the reciprocal rank of a list of ranked items
|
RR
|
{
"repo_name": "taolian/librec",
"path": "librec/src/main/java/librec/util/Measures.java",
"license": "gpl-3.0",
"size": 8835
}
|
[
"java.util.List"
] |
import java.util.List;
|
import java.util.*;
|
[
"java.util"
] |
java.util;
| 1,821,632
|
@ManyToOne(fetch = FetchType.EAGER)
@JoinColumn(name="invoice")
public Invoice getInvoice() {
return invoice;
}
|
@ManyToOne(fetch = FetchType.EAGER) @JoinColumn(name=STR) Invoice function() { return invoice; }
|
/**
* Gets the invoice.
*
* @return the invoice
*/
|
Gets the invoice
|
getInvoice
|
{
"repo_name": "Esleelkartea/aonGTA",
"path": "aongta_v1.0.0_src/Fuentes y JavaDoc/aon-finance/src/com/code/aon/finance/Finance.java",
"license": "gpl-2.0",
"size": 5400
}
|
[
"javax.persistence.FetchType",
"javax.persistence.JoinColumn",
"javax.persistence.ManyToOne"
] |
import javax.persistence.FetchType; import javax.persistence.JoinColumn; import javax.persistence.ManyToOne;
|
import javax.persistence.*;
|
[
"javax.persistence"
] |
javax.persistence;
| 630,856
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.